Add E126 pep8 rule
Updated rally code with E126 pep8 rule Closes-Bug:#1432078 Change-Id: Iefebdc6fdc4604869f26711240370b9e007e55df
This commit is contained in:
parent
1f29afc116
commit
2aa4dffa7f
@ -53,7 +53,7 @@ class Deployment(object):
|
||||
raise
|
||||
|
||||
deployer = deploy.EngineFactory.get_engine(
|
||||
deployment["config"]["type"], deployment)
|
||||
deployment["config"]["type"], deployment)
|
||||
try:
|
||||
deployer.validate()
|
||||
except jsonschema.ValidationError:
|
||||
@ -79,7 +79,7 @@ class Deployment(object):
|
||||
# is equal to "*->finished" or "deploy->inconsistent".
|
||||
deployment = objects.Deployment.get(deployment)
|
||||
deployer = deploy.EngineFactory.get_engine(
|
||||
deployment["config"]["type"], deployment)
|
||||
deployment["config"]["type"], deployment)
|
||||
|
||||
tempest.Tempest(deployment["uuid"]).uninstall()
|
||||
with deployer:
|
||||
@ -94,7 +94,7 @@ class Deployment(object):
|
||||
"""
|
||||
deployment = objects.Deployment.get(deployment)
|
||||
deployer = deploy.EngineFactory.get_engine(
|
||||
deployment["config"]["type"], deployment)
|
||||
deployment["config"]["type"], deployment)
|
||||
with deployer:
|
||||
deployer.make_cleanup()
|
||||
endpoints = deployer.make_deploy()
|
||||
|
@ -155,9 +155,9 @@ class SeekAndDestroy(object):
|
||||
|
||||
visited_tenants.add(user["tenant_id"])
|
||||
manager = self.manager_cls(
|
||||
admin=admin_client,
|
||||
user=self._get_cached_client(user),
|
||||
tenant_uuid=user["tenant_id"])
|
||||
admin=admin_client,
|
||||
user=self._get_cached_client(user),
|
||||
tenant_uuid=user["tenant_id"])
|
||||
|
||||
_publish(self.admin, user, manager)
|
||||
|
||||
@ -171,10 +171,10 @@ class SeekAndDestroy(object):
|
||||
admin, user, raw_resource = args
|
||||
|
||||
manager = self.manager_cls(
|
||||
resource=raw_resource,
|
||||
admin=self._get_cached_client(admin, cache=cache),
|
||||
user=self._get_cached_client(user, cache=cache),
|
||||
tenant_uuid=user and user["tenant_id"])
|
||||
resource=raw_resource,
|
||||
admin=self._get_cached_client(admin, cache=cache),
|
||||
user=self._get_cached_client(user, cache=cache),
|
||||
tenant_uuid=user and user["tenant_id"])
|
||||
|
||||
self._delete_single_resource(manager)
|
||||
|
||||
|
@ -126,6 +126,6 @@ class FlavorConfig(dict):
|
||||
:param extra_specs: is ignored
|
||||
"""
|
||||
super(FlavorConfig, self).__init__(
|
||||
name=name, ram=ram, vcpus=vcpus, disk=disk,
|
||||
swap=swap, ephemeral=ephemeral)
|
||||
name=name, ram=ram, vcpus=vcpus, disk=disk,
|
||||
swap=swap, ephemeral=ephemeral)
|
||||
self.__dict__.update(self)
|
||||
|
@ -76,11 +76,11 @@ def _prepare_open_secgroup(endpoint, secgroup_name):
|
||||
if not any(rule_match(new_rule, existing_rule) for existing_rule
|
||||
in rally_open.rules):
|
||||
nova.security_group_rules.create(
|
||||
rally_open.id,
|
||||
from_port=new_rule["from_port"],
|
||||
to_port=new_rule["to_port"],
|
||||
ip_protocol=new_rule["ip_protocol"],
|
||||
cidr=new_rule["ip_range"]["cidr"])
|
||||
rally_open.id,
|
||||
from_port=new_rule["from_port"],
|
||||
to_port=new_rule["to_port"],
|
||||
ip_protocol=new_rule["ip_protocol"],
|
||||
cidr=new_rule["ip_range"]["cidr"])
|
||||
|
||||
return rally_open.to_dict()
|
||||
|
||||
@ -121,4 +121,4 @@ class AllowSSH(base.Context):
|
||||
user["secgroup"]["name"]):
|
||||
clients = osclients.Clients(user["endpoint"])
|
||||
clients.nova().security_groups.get(
|
||||
user["secgroup"]["id"]).delete()
|
||||
user["secgroup"]["id"]).delete()
|
||||
|
@ -85,7 +85,7 @@ class ServerGenerator(base.Context):
|
||||
clients = osclients.Clients(user["endpoint"])
|
||||
nova_scenario = nova_utils.NovaScenario(clients=clients)
|
||||
server_name_prefix = nova_scenario._generate_random_name(
|
||||
prefix="ctx_rally_server_")
|
||||
prefix="ctx_rally_server_")
|
||||
|
||||
LOG.debug("Calling _boot_servers with server_name_prefix=%s "
|
||||
"image_id=%s flavor_id=%s servers_per_tenant=%s"
|
||||
|
@ -183,11 +183,11 @@ class UserGenerator(base.Context):
|
||||
"%s@email.me" % username,
|
||||
tenant_id, user_dom)
|
||||
user_endpoint = objects.Endpoint(
|
||||
client.auth_url, user.name, password,
|
||||
self.context["tenants"][tenant_id]["name"],
|
||||
consts.EndpointPermission.USER, client.region_name,
|
||||
project_domain_name=project_dom, user_domain_name=user_dom,
|
||||
endpoint_type=self.endpoint.endpoint_type)
|
||||
client.auth_url, user.name, password,
|
||||
self.context["tenants"][tenant_id]["name"],
|
||||
consts.EndpointPermission.USER, client.region_name,
|
||||
project_domain_name=project_dom, user_domain_name=user_dom,
|
||||
endpoint_type=self.endpoint.endpoint_type)
|
||||
users.append({"id": user.id,
|
||||
"endpoint": user_endpoint,
|
||||
"tenant_id": tenant_id})
|
||||
@ -241,8 +241,8 @@ class UserGenerator(base.Context):
|
||||
|
||||
if len(self.context["tenants"]) < self.config["tenants"]:
|
||||
raise exceptions.ContextSetupFailure(
|
||||
ctx_name=self.get_name(),
|
||||
msg=_("Failed to create the requested number of tenants."))
|
||||
ctx_name=self.get_name(),
|
||||
msg=_("Failed to create the requested number of tenants."))
|
||||
|
||||
users_num = self.config["users_per_tenant"] * self.config["tenants"]
|
||||
LOG.debug("Creating %(users)d users using %(threads)s threads" %
|
||||
@ -251,8 +251,8 @@ class UserGenerator(base.Context):
|
||||
|
||||
if len(self.context["users"]) < users_num:
|
||||
raise exceptions.ContextSetupFailure(
|
||||
ctx_name=self.get_name(),
|
||||
msg=_("Failed to create the requested number of users."))
|
||||
ctx_name=self.get_name(),
|
||||
msg=_("Failed to create the requested number of users."))
|
||||
|
||||
@rutils.log_task_wrapper(LOG.info, _("Exit context: `users`"))
|
||||
def cleanup(self):
|
||||
|
@ -63,7 +63,7 @@ class VolumeGenerator(base.Context):
|
||||
cinder_util = cinder_utils.CinderScenario(clients=clients)
|
||||
for i in range(volumes_per_tenant):
|
||||
rnd_name = scenario_base.Scenario._generate_random_name(
|
||||
prefix="ctx_rally_volume_")
|
||||
prefix="ctx_rally_volume_")
|
||||
vol = cinder_util._create_volume(size, display_name=rnd_name)
|
||||
self.context["tenants"][tenant_id]["volumes"].append(vol._info)
|
||||
|
||||
|
@ -90,20 +90,20 @@ def hvariety(data):
|
||||
raise ValueError("Cannot calculate number of histrogram bins "
|
||||
"for zero length array of data")
|
||||
return [
|
||||
{
|
||||
"method": "Square Root Choice",
|
||||
"number_of_bins": calculate_number_of_bins_sqrt(data),
|
||||
},
|
||||
{
|
||||
"method": "Sturges Formula",
|
||||
"number_of_bins": calculate_number_of_bins_sturges(data),
|
||||
},
|
||||
{
|
||||
"method": "Rice Rule",
|
||||
"number_of_bins": calculate_number_of_bins_rice(data),
|
||||
},
|
||||
{
|
||||
"method": "One Half",
|
||||
"number_of_bins": calculate_number_of_bins_half(data),
|
||||
}
|
||||
{
|
||||
"method": "Square Root Choice",
|
||||
"number_of_bins": calculate_number_of_bins_sqrt(data),
|
||||
},
|
||||
{
|
||||
"method": "Sturges Formula",
|
||||
"number_of_bins": calculate_number_of_bins_sturges(data),
|
||||
},
|
||||
{
|
||||
"method": "Rice Rule",
|
||||
"number_of_bins": calculate_number_of_bins_rice(data),
|
||||
},
|
||||
{
|
||||
"method": "One Half",
|
||||
"number_of_bins": calculate_number_of_bins_half(data),
|
||||
}
|
||||
]
|
||||
|
@ -28,7 +28,7 @@ def mean(values):
|
||||
"""
|
||||
if not values:
|
||||
raise exceptions.InvalidArgumentsException(
|
||||
"the list should be non-empty")
|
||||
"the list should be non-empty")
|
||||
return math.fsum(values) / len(values)
|
||||
|
||||
|
||||
|
@ -114,8 +114,8 @@ class Scenario(functional.FunctionalMixin):
|
||||
for func in dir(scenario) if Scenario.is_scenario(scenario, func)]
|
||||
for scenario in scenario_classes
|
||||
]
|
||||
benchmark_scenarios_flattened = list(itertools.chain.from_iterable(
|
||||
benchmark_scenarios))
|
||||
benchmark_scenarios_flattened = list(
|
||||
itertools.chain.from_iterable(benchmark_scenarios))
|
||||
return benchmark_scenarios_flattened
|
||||
|
||||
@staticmethod
|
||||
@ -222,7 +222,7 @@ class Scenario(functional.FunctionalMixin):
|
||||
"""
|
||||
if not 0 <= min_sleep <= max_sleep:
|
||||
raise exceptions.InvalidArgumentsException(
|
||||
"0 <= min_sleep <= max_sleep")
|
||||
"0 <= min_sleep <= max_sleep")
|
||||
|
||||
sleep_time = random.uniform(min_sleep, max_sleep)
|
||||
time.sleep(sleep_time)
|
||||
@ -295,7 +295,7 @@ class AtomicAction(utils.Timer):
|
||||
name_template = name + " (%i)"
|
||||
atomic_action_iteration = 2
|
||||
while self.scenario_instance._atomic_action_registered(
|
||||
name_template % atomic_action_iteration):
|
||||
name_template % atomic_action_iteration):
|
||||
atomic_action_iteration += 1
|
||||
return name_template % atomic_action_iteration
|
||||
|
||||
|
@ -177,7 +177,7 @@ class CeilometerScenario(base.Scenario):
|
||||
:returns: queried alarms
|
||||
"""
|
||||
return self.clients("ceilometer").query_alarms.query(
|
||||
filter, orderby, limit)
|
||||
filter, orderby, limit)
|
||||
|
||||
@base.atomic_action_timer("ceilometer.query_alarm_history")
|
||||
def _query_alarm_history(self, filter, orderby, limit):
|
||||
@ -192,7 +192,7 @@ class CeilometerScenario(base.Scenario):
|
||||
:returns: alarm history
|
||||
"""
|
||||
return self.clients("ceilometer").query_alarm_history.query(
|
||||
filter, orderby, limit)
|
||||
filter, orderby, limit)
|
||||
|
||||
@base.atomic_action_timer("ceilometer.create_sample")
|
||||
def _create_sample(self, counter_name, counter_type, counter_unit,
|
||||
@ -229,4 +229,4 @@ class CeilometerScenario(base.Scenario):
|
||||
:returns: queried samples
|
||||
"""
|
||||
return self.clients("ceilometer").query_samples.query(
|
||||
filter, orderby, limit)
|
||||
filter, orderby, limit)
|
||||
|
@ -43,7 +43,7 @@ class KeystoneScenario(base.Scenario):
|
||||
password = kwargs.pop("password", str(uuid.uuid4()))
|
||||
email = email or (name + "@rally.me")
|
||||
return self.admin_clients("keystone").users.create(
|
||||
name, password=password, email=email, **kwargs)
|
||||
name, password=password, email=email, **kwargs)
|
||||
|
||||
def _resource_delete(self, resource):
|
||||
""""Delete keystone resource."""
|
||||
@ -92,7 +92,7 @@ class KeystoneScenario(base.Scenario):
|
||||
password = name
|
||||
email = (name + "@rally.me")
|
||||
self.admin_clients("keystone").users.create(
|
||||
name, password=password, email=email, tenant_id=tenant.id)
|
||||
name, password=password, email=email, tenant_id=tenant.id)
|
||||
|
||||
@base.atomic_action_timer("keystone.create_role")
|
||||
def _role_create(self, name_length=5):
|
||||
|
@ -117,10 +117,10 @@ class NeutronNetworks(utils.NeutronScenario):
|
||||
:param subnets_per_network: int, number of subnets for one network
|
||||
"""
|
||||
network, subnets = self._create_network_and_subnets(
|
||||
network_create_args or {},
|
||||
subnet_create_args or {},
|
||||
subnets_per_network,
|
||||
subnet_cidr_start)
|
||||
network_create_args or {},
|
||||
subnet_create_args or {},
|
||||
subnets_per_network,
|
||||
subnet_cidr_start)
|
||||
|
||||
for subnet in subnets:
|
||||
self._update_subnet(subnet, subnet_update_args)
|
||||
@ -144,10 +144,10 @@ class NeutronNetworks(utils.NeutronScenario):
|
||||
:param subnets_per_network: int, number of subnets for one network
|
||||
"""
|
||||
network, subnets = self._create_network_and_subnets(
|
||||
network_create_args or {},
|
||||
subnet_create_args or {},
|
||||
subnets_per_network,
|
||||
subnet_cidr_start)
|
||||
network_create_args or {},
|
||||
subnet_create_args or {},
|
||||
subnets_per_network,
|
||||
subnet_cidr_start)
|
||||
|
||||
for subnet in subnets:
|
||||
self._delete_subnet(subnet)
|
||||
@ -174,10 +174,10 @@ class NeutronNetworks(utils.NeutronScenario):
|
||||
:param router_create_args: dict, POST /v2.0/routers request options
|
||||
"""
|
||||
network, subnets = self._create_network_and_subnets(
|
||||
network_create_args or {},
|
||||
subnet_create_args or {},
|
||||
subnets_per_network,
|
||||
subnet_cidr_start)
|
||||
network_create_args or {},
|
||||
subnet_create_args or {},
|
||||
subnets_per_network,
|
||||
subnet_cidr_start)
|
||||
|
||||
for subnet in subnets:
|
||||
router = self._create_router(router_create_args or {})
|
||||
@ -211,10 +211,10 @@ class NeutronNetworks(utils.NeutronScenario):
|
||||
:param router_create_args: dict, POST /v2.0/routers request options
|
||||
"""
|
||||
network, subnets = self._create_network_and_subnets(
|
||||
network_create_args or {},
|
||||
subnet_create_args or {},
|
||||
subnets_per_network,
|
||||
subnet_cidr_start)
|
||||
network_create_args or {},
|
||||
subnet_create_args or {},
|
||||
subnets_per_network,
|
||||
subnet_cidr_start)
|
||||
|
||||
for subnet in subnets:
|
||||
router = self._create_router(router_create_args or {})
|
||||
@ -244,10 +244,10 @@ class NeutronNetworks(utils.NeutronScenario):
|
||||
:param router_create_args: dict, POST /v2.0/routers request options
|
||||
"""
|
||||
network, subnets = self._create_network_and_subnets(
|
||||
network_create_args or {},
|
||||
subnet_create_args or {},
|
||||
subnets_per_network,
|
||||
subnet_cidr_start)
|
||||
network_create_args or {},
|
||||
subnet_create_args or {},
|
||||
subnets_per_network,
|
||||
subnet_cidr_start)
|
||||
|
||||
routers = []
|
||||
for subnet in subnets:
|
||||
@ -261,8 +261,8 @@ class NeutronNetworks(utils.NeutronScenario):
|
||||
router = routers[e]
|
||||
subnet = subnets[e]
|
||||
self.clients("neutron").remove_interface_router(
|
||||
router["router"]["id"],
|
||||
{"subnet_id": subnet["subnet"]["id"]})
|
||||
router["router"]["id"],
|
||||
{"subnet_id": subnet["subnet"]["id"]})
|
||||
self._delete_router(router)
|
||||
|
||||
@validation.number("ports_per_network", minval=1, integer_only=True)
|
||||
|
@ -57,7 +57,7 @@ class NeutronScenario(base.Scenario):
|
||||
:returns: updated neutron network dict
|
||||
"""
|
||||
suffix = network_update_args.get(
|
||||
"name", self._generate_random_name("_"))
|
||||
"name", self._generate_random_name("_"))
|
||||
admin_state_up = network_update_args.get("admin_state_up", True)
|
||||
body = {
|
||||
"network": {
|
||||
@ -116,7 +116,7 @@ class NeutronScenario(base.Scenario):
|
||||
:returns: updated neutron subnet dict
|
||||
"""
|
||||
suffix = subnet_update_args.get(
|
||||
"name", self._generate_random_name("_"))
|
||||
"name", self._generate_random_name("_"))
|
||||
enable_dhcp = subnet_update_args.get("enable_dhcp", True)
|
||||
body = {
|
||||
"subnet": {
|
||||
@ -182,7 +182,7 @@ class NeutronScenario(base.Scenario):
|
||||
:returns: updated neutron router dict
|
||||
"""
|
||||
suffix = router_update_args.get(
|
||||
"name", self._generate_random_name("_"))
|
||||
"name", self._generate_random_name("_"))
|
||||
admin_state = router_update_args.get("admin_state_up", True)
|
||||
body = {
|
||||
"router": {
|
||||
@ -224,7 +224,7 @@ class NeutronScenario(base.Scenario):
|
||||
:returns: updated neutron port dict
|
||||
"""
|
||||
suffix = port_update_args.get(
|
||||
"name", self._generate_random_name("_"))
|
||||
"name", self._generate_random_name("_"))
|
||||
admin_state = port_update_args.get("admin_state_up", True)
|
||||
device_owner = port_update_args.get("device_owner", "compute:nova")
|
||||
device_id = port_update_args.get("device_id", uid.generate_uuid())
|
||||
@ -286,4 +286,4 @@ class NeutronScenario(base.Scenario):
|
||||
:param router: dict, neutron router
|
||||
"""
|
||||
self.clients("neutron").remove_interface_router(
|
||||
router["id"], {"subnet_id": subnet["id"]})
|
||||
router["id"], {"subnet_id": subnet["id"]})
|
||||
|
@ -580,7 +580,7 @@ class NovaScenario(base.Scenario):
|
||||
update_resource=bench_utils.get_from_manager(),
|
||||
timeout=CONF.benchmark.nova_server_resize_revert_timeout,
|
||||
check_interval=(
|
||||
CONF.benchmark.nova_server_resize_revert_poll_interval)
|
||||
CONF.benchmark.nova_server_resize_revert_poll_interval)
|
||||
)
|
||||
|
||||
@base.atomic_action_timer("nova.attach_volume")
|
||||
@ -596,7 +596,7 @@ class NovaScenario(base.Scenario):
|
||||
update_resource=bench_utils.get_from_manager(),
|
||||
timeout=CONF.benchmark.nova_server_resize_revert_timeout,
|
||||
check_interval=(
|
||||
CONF.benchmark.nova_server_resize_revert_poll_interval)
|
||||
CONF.benchmark.nova_server_resize_revert_poll_interval)
|
||||
)
|
||||
|
||||
@base.atomic_action_timer("nova.detach_volume")
|
||||
@ -662,9 +662,9 @@ class NovaScenario(base.Scenario):
|
||||
break
|
||||
try:
|
||||
new_host = random.choice(
|
||||
[key for key, value in six.iteritems(az.hosts)
|
||||
if key != host and
|
||||
value["nova-compute"]["available"] is True])
|
||||
[key for key, value in six.iteritems(az.hosts)
|
||||
if key != host and
|
||||
value["nova-compute"]["available"] is True])
|
||||
return new_host
|
||||
except IndexError:
|
||||
raise exceptions.InvalidHostException(
|
||||
|
@ -118,7 +118,7 @@ class ActionBuilder(object):
|
||||
binding = self._bindings.get(action_key)
|
||||
dft_kwargs = dict(binding["kwargs"])
|
||||
dft_kwargs.update(kwargs or {})
|
||||
bound_actions.append(self._build(
|
||||
binding["action"], times,
|
||||
*(binding["args"] + args), **dft_kwargs))
|
||||
bound_actions.append(
|
||||
self._build(binding["action"], times,
|
||||
*(binding["args"] + args), **dft_kwargs))
|
||||
return bound_actions
|
||||
|
@ -179,7 +179,7 @@ def _get_validated_image(config, clients, param_name):
|
||||
if "regex" in image_args:
|
||||
match = re.match(image_args.get("regex"), image_ctx_name)
|
||||
if image_ctx_name == image_args.get("name") or (
|
||||
"regex" in image_args and match):
|
||||
"regex" in image_args and match):
|
||||
image = {
|
||||
"size": image_context.get("min_disk", 0),
|
||||
"min_ram": image_context.get("min_ram", 0),
|
||||
@ -301,7 +301,7 @@ def network_exists(config, clients, deployment, network_name):
|
||||
"Available networks: %(networks)s") % {
|
||||
"network": network,
|
||||
"networks": networks
|
||||
}
|
||||
}
|
||||
return ValidationResult(False, message)
|
||||
|
||||
|
||||
|
@ -190,7 +190,7 @@ class NeutronWrapper(NetworkWrapper):
|
||||
@property
|
||||
def external_networks(self):
|
||||
return self.client.list_networks(**{
|
||||
"router:external": True})["networks"]
|
||||
"router:external": True})["networks"]
|
||||
|
||||
def get_network(self, net_id=None, name=None):
|
||||
net = None
|
||||
|
@ -591,5 +591,5 @@ complete -F _rally rally
|
||||
args = " ".join(args_list)
|
||||
|
||||
completion.append(""" OPTS["{cat}_{cmd}"]="{args}"\n""".format(
|
||||
cat=category, cmd=command_name, args=args))
|
||||
cat=category, cmd=command_name, args=args))
|
||||
return bash_data % {"data": "".join(sorted(completion))}
|
||||
|
@ -260,7 +260,7 @@ def log_deprecated_args(message, rally_version, deprecated_args,
|
||||
def wrapper(*args, **kwargs):
|
||||
if (not once) or (not getattr(f, "_warned_dep_args", False)):
|
||||
deprecated = ", ".join([
|
||||
"`%s'" % x for x in deprecated_args if x in kwargs])
|
||||
"`%s'" % x for x in deprecated_args if x in kwargs])
|
||||
if deprecated:
|
||||
log_function(
|
||||
"%(msg)s (args %(args)s deprecated in Rally "
|
||||
|
@ -311,7 +311,7 @@ class Connection(object):
|
||||
return worker
|
||||
except db_exc.DBDuplicateEntry:
|
||||
raise exceptions.WorkerAlreadyRegistered(
|
||||
worker=values["hostname"])
|
||||
worker=values["hostname"])
|
||||
|
||||
def get_worker(self, hostname):
|
||||
try:
|
||||
|
@ -105,7 +105,7 @@ class VirshProvider(provider.ProviderFactory):
|
||||
subprocess.check_call(cmd, shell=True)
|
||||
|
||||
cmd = "virsh --connect=%s undefine %s --remove-all-storage" % (
|
||||
vconnection, vm_name)
|
||||
vconnection, vm_name)
|
||||
subprocess.check_call(cmd, shell=True)
|
||||
return True
|
||||
|
||||
|
@ -254,13 +254,13 @@ class Clients(object):
|
||||
auth_token = lambda: kc.auth_token
|
||||
|
||||
client = ceilometer.get_client(
|
||||
version,
|
||||
os_endpoint=metering_api_url,
|
||||
token=auth_token,
|
||||
timeout=CONF.openstack_client_http_timeout,
|
||||
insecure=self.endpoint.insecure,
|
||||
cacert=self.endpoint.cacert,
|
||||
**self._get_auth_info(project_name_key="tenant_name"))
|
||||
version,
|
||||
os_endpoint=metering_api_url,
|
||||
token=auth_token,
|
||||
timeout=CONF.openstack_client_http_timeout,
|
||||
insecure=self.endpoint.insecure,
|
||||
cacert=self.endpoint.cacert,
|
||||
**self._get_auth_info(project_name_key="tenant_name"))
|
||||
return client
|
||||
|
||||
@cached
|
||||
|
@ -171,7 +171,7 @@ class ConstantScenarioRunner(base.ScenarioRunner):
|
||||
|
||||
processes_to_start = min(max_cpu_used, times, concurrency)
|
||||
concurrency_per_worker, concurrency_overhead = divmod(
|
||||
concurrency, processes_to_start)
|
||||
concurrency, processes_to_start)
|
||||
|
||||
self._log_debug_info(times=times, concurrency=concurrency,
|
||||
timeout=timeout, max_cpu_used=max_cpu_used,
|
||||
@ -261,8 +261,8 @@ class ConstantForDurationScenarioRunner(base.ScenarioRunner):
|
||||
pool = multiprocessing.Pool(concurrency)
|
||||
|
||||
run_args = butils.infinite_run_args_generator(
|
||||
self._iter_scenario_args(cls, method, context, args,
|
||||
self.aborted))
|
||||
self._iter_scenario_args(cls, method, context, args,
|
||||
self.aborted))
|
||||
iter_result = pool.imap(base._run_scenario_once, run_args)
|
||||
|
||||
start = time.time()
|
||||
|
@ -171,8 +171,7 @@ class RPSScenarioRunner(base.ScenarioRunner):
|
||||
|
||||
# Determine concurrency per worker
|
||||
concurrency_per_worker, concurrency_overhead = divmod(
|
||||
self.config.get("max_concurrency", times),
|
||||
processes_to_start)
|
||||
self.config.get("max_concurrency", times), processes_to_start)
|
||||
|
||||
self._log_debug_info(times=times, timeout=timeout,
|
||||
max_cpu_used=max_cpu_used,
|
||||
|
@ -78,7 +78,7 @@ class UserCleanupTestCase(test.TestCase):
|
||||
mock_manager.list_resource_names.return_value = set(["a", "b", "c"])
|
||||
context.UserCleanup.validate(["a"])
|
||||
mock_manager.list_resource_names.assert_called_once_with(
|
||||
admin_required=False)
|
||||
admin_required=False)
|
||||
|
||||
@mock.patch("%s.manager" % BASE)
|
||||
def test_validate_no_such_cleanup(self, mock_manager):
|
||||
@ -86,7 +86,7 @@ class UserCleanupTestCase(test.TestCase):
|
||||
self.assertRaises(context.NoSuchCleanupResources,
|
||||
context.UserCleanup.validate, ["a", "b", "d"])
|
||||
mock_manager.list_resource_names.assert_called_once_with(
|
||||
admin_required=False)
|
||||
admin_required=False)
|
||||
|
||||
def test_validate_invalid_config(self):
|
||||
self.assertRaises(jsonschema.ValidationError,
|
||||
|
@ -91,8 +91,8 @@ class FlavorsGeneratorTestCase(test.TestCase):
|
||||
mock_osclients.assert_called_with(self.context["admin"]["endpoint"])
|
||||
|
||||
mock_flavor_create.assert_called_once_with(
|
||||
name="flavor_name", ram=2048, vcpus=3,
|
||||
disk=10, ephemeral=3, swap=5)
|
||||
name="flavor_name", ram=2048, vcpus=3,
|
||||
disk=10, ephemeral=3, swap=5)
|
||||
|
||||
@mock.patch("%s.flavors.osclients.Clients" % CTX)
|
||||
def test_cleanup(self, mock_osclients):
|
||||
|
@ -81,8 +81,8 @@ class RoleGeneratorTestCase(test.TestCase):
|
||||
{"id": "u2", "tenant_id": "t2"}]
|
||||
ctx._remove_role(mock.MagicMock(), role)
|
||||
calls = [
|
||||
mock.call("u1", role["id"], tenant="t1"),
|
||||
mock.call("u2", role["id"], tenant="t2"),
|
||||
mock.call("u1", role["id"], tenant="t1"),
|
||||
mock.call("u2", role["id"], tenant="t2"),
|
||||
]
|
||||
mock_keystone = mock_osclients.Clients().keystone()
|
||||
mock_keystone.roles.remove_user_role.assert_has_calls(calls)
|
||||
|
@ -226,8 +226,7 @@ class ScenarioRunnerTestCase(test.TestCase):
|
||||
config = {"a": 10}
|
||||
base.ScenarioRunner.validate(config)
|
||||
mock_validate.assert_called_once_with(
|
||||
config,
|
||||
serial.SerialScenarioRunner.CONFIG_SCHEMA)
|
||||
config, serial.SerialScenarioRunner.CONFIG_SCHEMA)
|
||||
|
||||
@mock.patch(BASE + "rutils.Timer.duration", return_value=10)
|
||||
def test_run(self, mock_duration):
|
||||
@ -302,7 +301,7 @@ class ScenarioRunnerTestCase(test.TestCase):
|
||||
processes = 10
|
||||
process_pool = collections.deque([process] * processes)
|
||||
mock_result_queue = mock.MagicMock(
|
||||
empty=mock.MagicMock(return_value=True))
|
||||
empty=mock.MagicMock(return_value=True))
|
||||
|
||||
runner = serial.SerialScenarioRunner(
|
||||
mock.MagicMock(),
|
||||
|
@ -187,7 +187,7 @@ class CinderServersTestCase(test.TestCase):
|
||||
|
||||
scenario._create_volume = mock.MagicMock(return_value=fake_volume)
|
||||
scenario._upload_volume_to_image = mock.MagicMock(
|
||||
return_value=fake_image)
|
||||
return_value=fake_image)
|
||||
scenario._delete_volume = mock.MagicMock()
|
||||
scenario._delete_image = mock.MagicMock()
|
||||
|
||||
|
@ -36,13 +36,9 @@ class KeystoneUtilsTestCase(test.TestCase):
|
||||
def test_is_temporary(self):
|
||||
prefix = utils.KeystoneScenario.RESOURCE_NAME_PREFIX
|
||||
tests = [
|
||||
(fakes.FakeResource(
|
||||
name=prefix + "abc"),
|
||||
True),
|
||||
(fakes.FakeResource(name=prefix + "abc"), True),
|
||||
(fakes.FakeResource(name="another"), False),
|
||||
(fakes.FakeResource(
|
||||
name=prefix[:-3] + "abc"),
|
||||
False)
|
||||
(fakes.FakeResource(name=prefix[:-3] + "abc"), False)
|
||||
]
|
||||
|
||||
for resource, is_valid in tests:
|
||||
@ -66,9 +62,9 @@ class KeystoneScenarioTestCase(test.TestCase):
|
||||
|
||||
self.assertEqual(user, result)
|
||||
fake_keystone.users.create.assert_called_once_with(
|
||||
mock_gen_name.return_value,
|
||||
password=mock_uuid4.return_value,
|
||||
email=mock_gen_name.return_value + "@rally.me")
|
||||
mock_gen_name.return_value,
|
||||
password=mock_uuid4.return_value,
|
||||
email=mock_gen_name.return_value + "@rally.me")
|
||||
mock_uuid4.assert_called_with()
|
||||
self._test_atomic_action_timer(scenario.atomic_actions(),
|
||||
"keystone.create_user")
|
||||
@ -161,8 +157,8 @@ class KeystoneScenarioTestCase(test.TestCase):
|
||||
scenario._users_create(tenant, users_per_tenant=1, name_length=10)
|
||||
|
||||
fake_keystone.users.create.assert_called_once_with(
|
||||
name, password=name, email=name + "@rally.me",
|
||||
tenant_id=tenant.id)
|
||||
name, password=name, email=name + "@rally.me",
|
||||
tenant_id=tenant.id)
|
||||
self._test_atomic_action_timer(scenario.atomic_actions(),
|
||||
"keystone.create_users")
|
||||
|
||||
|
@ -47,11 +47,11 @@ class NeutronNetworksTestCase(test.TestCase):
|
||||
|
||||
@mock.patch(NEUTRON_NETWORKS + "._update_network")
|
||||
@mock.patch(NEUTRON_NETWORKS + "._create_network", return_value={
|
||||
"network": {
|
||||
"id": "network-id",
|
||||
"name": "network-name",
|
||||
"admin_state_up": False
|
||||
}
|
||||
"network": {
|
||||
"id": "network-id",
|
||||
"name": "network-name",
|
||||
"admin_state_up": False
|
||||
}
|
||||
})
|
||||
def test_create_and_update_networks(self,
|
||||
mock_create_network,
|
||||
@ -99,7 +99,7 @@ class NeutronNetworksTestCase(test.TestCase):
|
||||
# Explict network name is specified
|
||||
network_create_args = {"name": "given-name"}
|
||||
neutron_scenario.create_and_delete_networks(
|
||||
network_create_args=network_create_args)
|
||||
network_create_args=network_create_args)
|
||||
mock_create.assert_called_once_with(network_create_args)
|
||||
self.assertEqual(1, mock_delete.call_count)
|
||||
|
||||
@ -540,17 +540,14 @@ class NeutronNetworksTestCase(test.TestCase):
|
||||
@mock.patch(NEUTRON_NETWORKS + "._generate_random_name")
|
||||
@mock.patch(NEUTRON_NETWORKS + "._update_port")
|
||||
@mock.patch(NEUTRON_NETWORKS + "._create_port", return_value={
|
||||
"port": {
|
||||
"name": "port-name",
|
||||
"id": "port-id",
|
||||
"admin_state_up": True
|
||||
}
|
||||
"port": {
|
||||
"name": "port-name",
|
||||
"id": "port-id",
|
||||
"admin_state_up": True
|
||||
}
|
||||
})
|
||||
@mock.patch(NEUTRON_NETWORKS + "._create_network", return_value={
|
||||
"network": {
|
||||
"id": "fake-id"
|
||||
}
|
||||
})
|
||||
"network": {"id": "fake-id"}})
|
||||
def test_create_and_update_ports(self,
|
||||
mock_create_network,
|
||||
mock_create_port,
|
||||
|
@ -84,7 +84,7 @@ class NeutronScenarioTestCase(test.TestCase):
|
||||
|
||||
result_network = scenario._update_network(network, network_update_args)
|
||||
mock_clients("neutron").update_network.assert_called_once_with(
|
||||
network["network"]["id"], expected_network)
|
||||
network["network"]["id"], expected_network)
|
||||
self.assertEqual(result_network, expected_network)
|
||||
self._test_atomic_action_timer(scenario.atomic_actions(),
|
||||
"neutron.update_network")
|
||||
@ -163,7 +163,7 @@ class NeutronScenarioTestCase(test.TestCase):
|
||||
|
||||
result_subnet = scenario._update_subnet(subnet, subnet_update_args)
|
||||
mock_clients("neutron").update_subnet.assert_called_once_with(
|
||||
subnet["subnet"]["id"], expected_subnet)
|
||||
subnet["subnet"]["id"], expected_subnet)
|
||||
self.assertEqual(result_subnet, expected_subnet)
|
||||
self._test_atomic_action_timer(scenario.atomic_actions(),
|
||||
"neutron.update_subnet")
|
||||
@ -237,7 +237,7 @@ class NeutronScenarioTestCase(test.TestCase):
|
||||
|
||||
result_router = scenario._update_router(router, router_update_args)
|
||||
mock_clients("neutron").update_router.assert_called_once_with(
|
||||
router["router"]["id"], expected_router)
|
||||
router["router"]["id"], expected_router)
|
||||
self.assertEqual(result_router, expected_router)
|
||||
self._test_atomic_action_timer(scenario.atomic_actions(),
|
||||
"neutron.update_router")
|
||||
@ -248,7 +248,7 @@ class NeutronScenarioTestCase(test.TestCase):
|
||||
router = scenario._create_router({})
|
||||
scenario._delete_router(router)
|
||||
mock_clients("neutron").delete_router.assert_called_once_with(
|
||||
router["router"]["id"])
|
||||
router["router"]["id"])
|
||||
self._test_atomic_action_timer(scenario.atomic_actions(),
|
||||
"neutron.delete_router")
|
||||
|
||||
@ -262,7 +262,7 @@ class NeutronScenarioTestCase(test.TestCase):
|
||||
scenario._remove_interface_router(subnet, router)
|
||||
mock_remove_router = mock_clients("neutron").remove_interface_router
|
||||
mock_remove_router.assert_called_once_with(
|
||||
router["id"], {"subnet_id": subnet["id"]})
|
||||
router["id"], {"subnet_id": subnet["id"]})
|
||||
self._test_atomic_action_timer(scenario.atomic_actions(),
|
||||
"neutron.remove_interface_router")
|
||||
|
||||
@ -341,7 +341,7 @@ class NeutronScenarioTestCase(test.TestCase):
|
||||
|
||||
result_port = scenario._update_port(port, port_update_args)
|
||||
mock_clients("neutron").update_port.assert_called_once_with(
|
||||
port["port"]["id"], expected_port)
|
||||
port["port"]["id"], expected_port)
|
||||
self.assertEqual(result_port, expected_port)
|
||||
self._test_atomic_action_timer(scenario.atomic_actions(),
|
||||
"neutron.update_port")
|
||||
|
@ -415,7 +415,7 @@ class NovaServersTestCase(test.TestCase):
|
||||
scenario._boot_server = mock.MagicMock(return_value=fake_server)
|
||||
scenario.sleep_between = mock.MagicMock()
|
||||
scenario._find_host_to_migrate = mock.MagicMock(
|
||||
return_value="host_name")
|
||||
return_value="host_name")
|
||||
scenario._live_migrate = mock.MagicMock()
|
||||
scenario._delete_server = mock.MagicMock()
|
||||
|
||||
@ -442,7 +442,7 @@ class NovaServersTestCase(test.TestCase):
|
||||
scenario._boot_server = mock.MagicMock(return_value=fake_server)
|
||||
scenario.sleep_between = mock.MagicMock()
|
||||
scenario._find_host_to_migrate = mock.MagicMock(
|
||||
return_value="host_name")
|
||||
return_value="host_name")
|
||||
scenario._live_migrate = mock.MagicMock()
|
||||
scenario._delete_server = mock.MagicMock()
|
||||
|
||||
@ -484,7 +484,7 @@ class NovaServersTestCase(test.TestCase):
|
||||
scenario.sleep_between = mock.MagicMock()
|
||||
|
||||
scenario._find_host_to_migrate = mock.MagicMock(
|
||||
return_value="host_name")
|
||||
return_value="host_name")
|
||||
scenario._live_migrate = mock.MagicMock()
|
||||
|
||||
scenario._boot_server = mock.MagicMock(return_value=fake_server)
|
||||
|
@ -126,9 +126,9 @@ class NovaScenarioTestCase(test.TestCase):
|
||||
{"id": "bar_id", "external": False}]
|
||||
mock_clients("nova").networks.list.return_value = networks
|
||||
nova_scenario = utils.NovaScenario(context={
|
||||
"iteration": 3,
|
||||
"config": {"users": {"tenants": 2}},
|
||||
"tenant": {"networks": networks}})
|
||||
"iteration": 3,
|
||||
"config": {"users": {"tenants": 2}},
|
||||
"tenant": {"networks": networks}})
|
||||
return_server = nova_scenario._boot_server("image_id",
|
||||
"flavor_id",
|
||||
auto_assign_nic=True)
|
||||
@ -587,7 +587,7 @@ class NovaScenarioTestCase(test.TestCase):
|
||||
nova_scenario = utils.NovaScenario(admin_clients=fakes.FakeClients())
|
||||
|
||||
self.assertIn(
|
||||
nova_scenario._find_host_to_migrate(fake_server), ["b1", "b3"])
|
||||
nova_scenario._find_host_to_migrate(fake_server), ["b1", "b3"])
|
||||
|
||||
@mock.patch(NOVA_UTILS + ".NovaScenario.clients")
|
||||
def test__migrate_server(self, mock_clients):
|
||||
|
@ -29,14 +29,14 @@ class QuotasScenarioTestCase(test.TestCase):
|
||||
def test__update_quotas(self):
|
||||
tenant_id = "fake_tenant"
|
||||
quotas = {
|
||||
"metadata_items": 10,
|
||||
"key_pairs": 10,
|
||||
"injected_file_content_bytes": 1024,
|
||||
"injected_file_path_bytes": 1024,
|
||||
"ram": 5120,
|
||||
"instances": 10,
|
||||
"injected_files": 10,
|
||||
"cores": 10,
|
||||
"metadata_items": 10,
|
||||
"key_pairs": 10,
|
||||
"injected_file_content_bytes": 1024,
|
||||
"injected_file_path_bytes": 1024,
|
||||
"ram": 5120,
|
||||
"instances": 10,
|
||||
"injected_files": 10,
|
||||
"cores": 10,
|
||||
}
|
||||
fake_nova = fakes.FakeNovaClient()
|
||||
fake_nova.quotas.update = mock.MagicMock(return_value=quotas)
|
||||
@ -55,14 +55,14 @@ class QuotasScenarioTestCase(test.TestCase):
|
||||
def test__update_quotas_fn(self):
|
||||
tenant_id = "fake_tenant"
|
||||
quotas = {
|
||||
"metadata_items": 10,
|
||||
"key_pairs": 10,
|
||||
"injected_file_content_bytes": 1024,
|
||||
"injected_file_path_bytes": 1024,
|
||||
"ram": 5120,
|
||||
"instances": 10,
|
||||
"injected_files": 10,
|
||||
"cores": 10,
|
||||
"metadata_items": 10,
|
||||
"key_pairs": 10,
|
||||
"injected_file_content_bytes": 1024,
|
||||
"injected_file_path_bytes": 1024,
|
||||
"ram": 5120,
|
||||
"instances": 10,
|
||||
"injected_files": 10,
|
||||
"cores": 10,
|
||||
}
|
||||
fake_nova = fakes.FakeNovaClient()
|
||||
fake_nova.quotas.update = mock.MagicMock(return_value=quotas)
|
||||
|
@ -44,12 +44,12 @@ class AuthenticateTestCase(test.TestCase):
|
||||
scenario = authenticate.Authenticate(admin_clients=mock_admin_clients,
|
||||
clients=mock_users_clients)
|
||||
scenario._clients.glance.images.list = mock.MagicMock(
|
||||
return_value=images_list)
|
||||
return_value=images_list)
|
||||
image_name = "__intentionally_non_existent_image___"
|
||||
with base.AtomicAction(scenario, "authenticate.validate_glance"):
|
||||
scenario.validate_glance(5)
|
||||
scenario._clients.glance().images.list.assert_called_with(
|
||||
name=image_name)
|
||||
name=image_name)
|
||||
self.assertEqual(scenario._clients.glance().images.list.call_count, 5)
|
||||
|
||||
@mock.patch("rally.osclients")
|
||||
@ -62,7 +62,7 @@ class AuthenticateTestCase(test.TestCase):
|
||||
scenario = authenticate.Authenticate(admin_clients=mock_admin_clients,
|
||||
clients=mock_users_clients)
|
||||
scenario._clients.nova.flavors.list = mock.MagicMock(
|
||||
return_value=flavors_list)
|
||||
return_value=flavors_list)
|
||||
with base.AtomicAction(scenario, "authenticate.validate_nova"):
|
||||
scenario.validate_nova(5)
|
||||
self.assertEqual(scenario._clients.nova().flavors.list.call_count, 5)
|
||||
@ -77,7 +77,7 @@ class AuthenticateTestCase(test.TestCase):
|
||||
scenario = authenticate.Authenticate(admin_clients=mock_admin_clients,
|
||||
clients=mock_users_clients)
|
||||
scenario._clients.cinder.volume_types.list = mock.MagicMock(
|
||||
return_value=volume_types_list)
|
||||
return_value=volume_types_list)
|
||||
with base.AtomicAction(scenario, "authenticate.validate_cinder"):
|
||||
scenario.validate_cinder(5)
|
||||
self.assertEqual(scenario._clients.cinder().volume_types.
|
||||
@ -107,7 +107,7 @@ class AuthenticateTestCase(test.TestCase):
|
||||
scenario = authenticate.Authenticate(admin_clients=mock_admin_clients,
|
||||
clients=mock_users_clients)
|
||||
scenario._clients.heat.stacks.list = mock.MagicMock(
|
||||
return_value=stacks_list)
|
||||
return_value=stacks_list)
|
||||
with base.AtomicAction(scenario, "authenticate.validate_heat"):
|
||||
scenario.validate_heat(5)
|
||||
scenario._clients.heat().stacks.list.assert_called_with(limit=0)
|
||||
|
@ -126,8 +126,8 @@ class VMScenarioTestCase(test.TestCase):
|
||||
self.assertTrue(vm_scenario._ping_ip_address(host_ip))
|
||||
|
||||
mock_subprocess.assert_called_once_with(
|
||||
["ping", "-c1", "-w1", host_ip],
|
||||
stderr=subprocess.PIPE, stdout=subprocess.PIPE)
|
||||
["ping", "-c1", "-w1", host_ip],
|
||||
stderr=subprocess.PIPE, stdout=subprocess.PIPE)
|
||||
ping_process.wait.assert_called_once_with()
|
||||
|
||||
@mock.patch(VMTASKS_UTILS + ".sys")
|
||||
@ -143,8 +143,8 @@ class VMScenarioTestCase(test.TestCase):
|
||||
self.assertTrue(vm_scenario._ping_ip_address(host_ip))
|
||||
|
||||
mock_subprocess.assert_called_once_with(
|
||||
["ping6", "-c1", "-w1", host_ip],
|
||||
stderr=subprocess.PIPE, stdout=subprocess.PIPE)
|
||||
["ping6", "-c1", "-w1", host_ip],
|
||||
stderr=subprocess.PIPE, stdout=subprocess.PIPE)
|
||||
ping_process.wait.assert_called_once_with()
|
||||
|
||||
@mock.patch(VMTASKS_UTILS + ".sys")
|
||||
@ -160,8 +160,8 @@ class VMScenarioTestCase(test.TestCase):
|
||||
self.assertTrue(vm_scenario._ping_ip_address(host_ip))
|
||||
|
||||
mock_subprocess.assert_called_once_with(
|
||||
["ping", "-c1", host_ip],
|
||||
stderr=subprocess.PIPE, stdout=subprocess.PIPE)
|
||||
["ping", "-c1", host_ip],
|
||||
stderr=subprocess.PIPE, stdout=subprocess.PIPE)
|
||||
ping_process.wait.assert_called_once_with()
|
||||
|
||||
@mock.patch(VMTASKS_UTILS + ".sys")
|
||||
@ -177,8 +177,8 @@ class VMScenarioTestCase(test.TestCase):
|
||||
self.assertTrue(vm_scenario._ping_ip_address(host_ip))
|
||||
|
||||
mock_subprocess.assert_called_once_with(
|
||||
["ping6", "-c1", host_ip],
|
||||
stderr=subprocess.PIPE, stdout=subprocess.PIPE)
|
||||
["ping6", "-c1", host_ip],
|
||||
stderr=subprocess.PIPE, stdout=subprocess.PIPE)
|
||||
ping_process.wait.assert_called_once_with()
|
||||
|
||||
def get_scenario(self):
|
||||
@ -203,8 +203,8 @@ class VMScenarioTestCase(test.TestCase):
|
||||
scenario._boot_server_with_fip,
|
||||
"foo_image", "foo_flavor", foo_arg="foo_value")
|
||||
scenario._boot_server.assert_called_once_with(
|
||||
"foo_image", "foo_flavor",
|
||||
foo_arg="foo_value", auto_assign_nic=True)
|
||||
"foo_image", "foo_flavor",
|
||||
foo_arg="foo_value", auto_assign_nic=True)
|
||||
|
||||
def test__boot_server_with_fixed_ip(self):
|
||||
scenario, server = self.get_scenario()
|
||||
|
@ -37,14 +37,14 @@ class VMTasksTestCase(test.TestCase):
|
||||
|
||||
def test_boot_runcommand_delete(self):
|
||||
self.scenario.boot_runcommand_delete(
|
||||
"foo_image", "foo_flavor", "foo_script",
|
||||
"foo_interpreter", "foo_username",
|
||||
password="foo_password",
|
||||
use_floating_ip="use_fip",
|
||||
floating_network="ext_network",
|
||||
force_delete="foo_force",
|
||||
volume_args={"size": 16},
|
||||
foo_arg="foo_value")
|
||||
"foo_image", "foo_flavor", "foo_script",
|
||||
"foo_interpreter", "foo_username",
|
||||
password="foo_password",
|
||||
use_floating_ip="use_fip",
|
||||
floating_network="ext_network",
|
||||
force_delete="foo_force",
|
||||
volume_args={"size": 16},
|
||||
foo_arg="foo_value")
|
||||
|
||||
self.scenario._create_volume.assert_called_once_with(
|
||||
16, imageRef=None)
|
||||
|
@ -143,7 +143,7 @@ class BenchmarkEngineTestCase(test.TestCase):
|
||||
eng = engine.BenchmarkEngine(mock.MagicMock(), mock.MagicMock())
|
||||
|
||||
mock_runner.validate = mock.MagicMock(
|
||||
side_effect=jsonschema.ValidationError("a"))
|
||||
side_effect=jsonschema.ValidationError("a"))
|
||||
self.assertRaises(exceptions.InvalidBenchmarkConfig,
|
||||
eng._validate_config_syntax, config)
|
||||
|
||||
@ -155,7 +155,7 @@ class BenchmarkEngineTestCase(test.TestCase):
|
||||
eng = engine.BenchmarkEngine(mock.MagicMock(), mock.MagicMock())
|
||||
|
||||
mock_context.validate = mock.MagicMock(
|
||||
side_effect=jsonschema.ValidationError("a"))
|
||||
side_effect=jsonschema.ValidationError("a"))
|
||||
self.assertRaises(exceptions.InvalidBenchmarkConfig,
|
||||
eng._validate_config_syntax, config)
|
||||
|
||||
|
@ -38,21 +38,18 @@ class FlavorResourceTypeTestCase(test.TestCase):
|
||||
def test_transform_by_id(self):
|
||||
resource_config = {"id": "42"}
|
||||
flavor_id = types.FlavorResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(flavor_id, "42")
|
||||
|
||||
def test_transform_by_id_from_base_class(self):
|
||||
resource_config = {}
|
||||
types.ResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
|
||||
def test_transform_by_name(self):
|
||||
resource_config = {"name": "m1.nano"}
|
||||
flavor_id = types.FlavorResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(flavor_id, "42")
|
||||
|
||||
def test_transform_by_name_no_match(self):
|
||||
@ -70,8 +67,7 @@ class FlavorResourceTypeTestCase(test.TestCase):
|
||||
def test_transform_by_regex(self):
|
||||
resource_config = {"regex": "m(1|2)\.nano"}
|
||||
flavor_id = types.FlavorResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(flavor_id, "42")
|
||||
|
||||
def test_transform_by_regex_multiple_match(self):
|
||||
@ -104,15 +100,13 @@ class EC2FlavorResourceTypeTestCase(test.TestCase):
|
||||
def test_transform_by_name(self):
|
||||
resource_config = {"name": "m1.nano"}
|
||||
flavor_name = types.EC2FlavorResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(flavor_name, "m1.nano")
|
||||
|
||||
def test_transform_by_id(self):
|
||||
resource_config = {"id": "2"}
|
||||
flavor_name = types.EC2FlavorResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(flavor_name, "m1.nano")
|
||||
|
||||
def test_transform_by_id_no_match(self):
|
||||
@ -147,15 +141,13 @@ class ImageResourceTypeTestCase(test.TestCase):
|
||||
def test_transform_by_id(self):
|
||||
resource_config = {"id": "100"}
|
||||
image_id = types.ImageResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(image_id, "100")
|
||||
|
||||
def test_transform_by_name(self):
|
||||
resource_config = {"name": "^cirros-0.3.1-uec$"}
|
||||
image_id = types.ImageResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(image_id, "100")
|
||||
|
||||
def test_transform_by_name_no_match(self):
|
||||
@ -173,8 +165,7 @@ class ImageResourceTypeTestCase(test.TestCase):
|
||||
def test_transform_by_regex(self):
|
||||
resource_config = {"regex": "-uec$"}
|
||||
image_id = types.ImageResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(image_id, "100")
|
||||
|
||||
def test_transform_by_regex_match_multiple(self):
|
||||
@ -220,15 +211,13 @@ class EC2ImageResourceTypeTestCase(test.TestCase):
|
||||
def test_transform_by_name(self):
|
||||
resource_config = {"name": "^cirros-0.3.1-uec$"}
|
||||
ec2_image_id = types.EC2ImageResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(ec2_image_id, "200")
|
||||
|
||||
def test_transform_by_id(self):
|
||||
resource_config = {"id": "100"}
|
||||
ec2_image_id = types.EC2ImageResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(ec2_image_id, "200")
|
||||
|
||||
def test_transform_by_id_no_match(self):
|
||||
@ -258,8 +247,7 @@ class EC2ImageResourceTypeTestCase(test.TestCase):
|
||||
def test_transform_by_regex(self):
|
||||
resource_config = {"regex": "-uec$"}
|
||||
ec2_image_id = types.EC2ImageResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(ec2_image_id, "200")
|
||||
|
||||
def test_transform_by_regex_match_multiple(self):
|
||||
@ -286,15 +274,13 @@ class VolumeTypeResourceTypeTestCase(test.TestCase):
|
||||
def test_transform_by_id(self):
|
||||
resource_config = {"id": 100}
|
||||
volumetype_id = types.VolumeTypeResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(volumetype_id, 100)
|
||||
|
||||
def test_transform_by_name(self):
|
||||
resource_config = {"name": "lvmdriver-1"}
|
||||
volumetype_id = types.VolumeTypeResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(volumetype_id, 100)
|
||||
|
||||
def test_transform_by_name_no_match(self):
|
||||
@ -306,8 +292,7 @@ class VolumeTypeResourceTypeTestCase(test.TestCase):
|
||||
def test_transform_by_regex(self):
|
||||
resource_config = {"regex": "^lvm.*-1"}
|
||||
volumetype_id = types.VolumeTypeResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(volumetype_id, 100)
|
||||
|
||||
def test_transform_by_regex_no_match(self):
|
||||
@ -331,15 +316,13 @@ class NeutronNetworkResourceTypeTestCase(test.TestCase):
|
||||
def test_transform_by_id(self):
|
||||
resource_config = {"id": self.net1_id}
|
||||
network_id = types.NeutronNetworkResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(network_id, self.net1_id)
|
||||
|
||||
def test_transform_by_name(self):
|
||||
resource_config = {"name": "net1"}
|
||||
network_id = types.NeutronNetworkResourceType.transform(
|
||||
clients=self.clients,
|
||||
resource_config=resource_config)
|
||||
clients=self.clients, resource_config=resource_config)
|
||||
self.assertEqual(network_id, self.net1_id)
|
||||
|
||||
def test_transform_by_name_no_match(self):
|
||||
@ -387,8 +370,7 @@ class FileTypeTestCase(test.TestCase):
|
||||
def test_transform_by_path(self, mock_open):
|
||||
resource_config = "file.yaml"
|
||||
file_context = types.FileType.transform(
|
||||
clients=None,
|
||||
resource_config=resource_config)
|
||||
clients=None, resource_config=resource_config)
|
||||
self.assertEqual(file_context, "file_context")
|
||||
|
||||
@mock.patch("rally.benchmark.types.open",
|
||||
|
@ -630,8 +630,8 @@ class ValidatorsTestCase(test.TestCase):
|
||||
|
||||
def test_required_cinder_services(self):
|
||||
validator = self._unwrap_validator(
|
||||
validation.required_cinder_services,
|
||||
service_name=six.text_type("cinder-service"))
|
||||
validation.required_cinder_services,
|
||||
service_name=six.text_type("cinder-service"))
|
||||
|
||||
with mock.patch.object(rally.osclients.Clients, "cinder") as client:
|
||||
fake_service = mock.Mock(binary="cinder-service", state="up")
|
||||
|
@ -115,11 +115,11 @@ class DeploymentCommandsTestCase(test.TestCase):
|
||||
current_deployment_id = "26a3ce76-0efa-40e4-86e5-514574bd1ff6"
|
||||
mock_default.return_value = current_deployment_id
|
||||
fake_deployment_list = [
|
||||
{"uuid": "fa34aea2-ae2e-4cf7-a072-b08d67466e3e",
|
||||
"created_at": "03-12-2014",
|
||||
"name": "dep1",
|
||||
"status": "deploy->started",
|
||||
"active": "False"}]
|
||||
{"uuid": "fa34aea2-ae2e-4cf7-a072-b08d67466e3e",
|
||||
"created_at": "03-12-2014",
|
||||
"name": "dep1",
|
||||
"status": "deploy->started",
|
||||
"active": "False"}]
|
||||
|
||||
mock_deployments.return_value = fake_deployment_list
|
||||
self.deployment.list()
|
||||
@ -220,12 +220,12 @@ class DeploymentCommandsTestCase(test.TestCase):
|
||||
deployment_id = mock_deployment.return_value["uuid"]
|
||||
|
||||
mock_deployment.return_value["admin"] = {
|
||||
"auth_url": "fake_auth_url",
|
||||
"username": "fake_username",
|
||||
"password": "fake_password",
|
||||
"tenant_name": "fake_tenant_name",
|
||||
"endpoint": "fake_endpoint",
|
||||
"region_name": None}
|
||||
"auth_url": "fake_auth_url",
|
||||
"username": "fake_username",
|
||||
"password": "fake_password",
|
||||
"tenant_name": "fake_tenant_name",
|
||||
"endpoint": "fake_endpoint",
|
||||
"region_name": None}
|
||||
|
||||
with mock.patch("rally.cmd.commands.deployment.open", mock.mock_open(),
|
||||
create=True) as mock_file:
|
||||
|
@ -56,4 +56,4 @@ class TempestCommandsTestCase(test.TestCase):
|
||||
deployment_uuid = "deployment_uuid"
|
||||
self.tempest_commands.install(deployment_uuid)
|
||||
mock_api.Verification.install_tempest.assert_called_once_with(
|
||||
deployment_uuid, None)
|
||||
deployment_uuid, None)
|
||||
|
@ -81,9 +81,8 @@ class TasksTestCase(test.DBTestCase):
|
||||
FINISHED = consts.TaskStatus.FINISHED
|
||||
task_finished = sorted(self._create_task(
|
||||
{"status": FINISHED,
|
||||
"deployment_uuid": self.deploy["uuid"]})[
|
||||
"uuid"]
|
||||
for i in moves.range(3))
|
||||
"deployment_uuid": self.deploy["uuid"]}
|
||||
)["uuid"] for i in moves.range(3))
|
||||
|
||||
task_all = sorted(task_init + task_finished)
|
||||
|
||||
|
@ -90,9 +90,9 @@ class DevstackEngineTestCase(test.TestCase):
|
||||
m_endpoint.assert_called_once_with("http://host:5000/v2.0/", "admin",
|
||||
"secret", "admin", "admin")
|
||||
m_d.add_resource.assert_called_once_with(
|
||||
info="fake_credentials",
|
||||
provider_name="DevstackEngine",
|
||||
type="credentials")
|
||||
info="fake_credentials",
|
||||
provider_name="DevstackEngine",
|
||||
type="credentials")
|
||||
repo = "https://git.openstack.org/openstack-dev/devstack"
|
||||
cmd = "/bin/sh -e -s %s master" % repo
|
||||
server.ssh.run.assert_called_once_with(cmd, stdin="fake_script")
|
||||
|
@ -76,7 +76,7 @@ class OpenStackProviderTestCase(test.TestCase):
|
||||
|
||||
self.nova_client = mock.MagicMock()
|
||||
self.nova_client.servers.create = mock.MagicMock(
|
||||
return_value=self.instance)
|
||||
return_value=self.instance)
|
||||
|
||||
self.clients.nova = mock.MagicMock(return_value=self.nova_client)
|
||||
|
||||
|
@ -47,7 +47,7 @@ class VirshProviderTestCase(test.TestCase):
|
||||
mock_ipaddress.return_value = "10.0.0.2"
|
||||
server = self.provider.create_vm("name")
|
||||
script_path = "%(virsh_path)s/virsh/get_domain_ip.sh" % dict(
|
||||
virsh_path=os.path.split(virsh.__file__)[0])
|
||||
virsh_path=os.path.split(virsh.__file__)[0])
|
||||
mock_subp.assert_has_calls([
|
||||
mock.call.check_call("virt-clone --connect=qemu+ssh://user@host/"
|
||||
"system -o prefix -n name --auto-clone",
|
||||
|
@ -1150,12 +1150,11 @@ class FakeNeutronClient(object):
|
||||
return {"router": router}
|
||||
|
||||
def create_subnet(self, data):
|
||||
subnet = setup_dict(data["subnet"],
|
||||
required=["network_id", "cidr", "ip_version"],
|
||||
defaults={
|
||||
"name": generate_name("subnet_"),
|
||||
"dns_nameservers": ["8.8.8.8", "8.8.4.4"]
|
||||
})
|
||||
subnet = setup_dict(
|
||||
data["subnet"],
|
||||
required=["network_id", "cidr", "ip_version"],
|
||||
defaults={"name": generate_name("subnet_"),
|
||||
"dns_nameservers": ["8.8.8.8", "8.8.4.4"]})
|
||||
if subnet["network_id"] not in self.__networks:
|
||||
raise neutron_exceptions.NeutronClientException
|
||||
|
||||
|
@ -38,7 +38,7 @@ class TaskTestCase(test.TestCase):
|
||||
mock_create.return_value = self.task
|
||||
task = objects.Task(status=consts.TaskStatus.FAILED)
|
||||
mock_create.assert_called_once_with({
|
||||
"status": consts.TaskStatus.FAILED})
|
||||
"status": consts.TaskStatus.FAILED})
|
||||
self.assertEqual(task["uuid"], self.task["uuid"])
|
||||
|
||||
@mock.patch("rally.objects.task.db.task_create")
|
||||
|
@ -268,7 +268,7 @@ class ConstantForDurationScenarioRunnerTestCase(test.TestCase):
|
||||
|
||||
def test_run_scenario_constantly_for_duration(self):
|
||||
runner = constant.ConstantForDurationScenarioRunner(
|
||||
None, self.config)
|
||||
None, self.config)
|
||||
|
||||
runner._run_scenario(fakes.FakeScenario, "do_it",
|
||||
self.context, self.args)
|
||||
@ -280,7 +280,7 @@ class ConstantForDurationScenarioRunnerTestCase(test.TestCase):
|
||||
|
||||
def test_run_scenario_constantly_for_duration_exception(self):
|
||||
runner = constant.ConstantForDurationScenarioRunner(
|
||||
None, self.config)
|
||||
None, self.config)
|
||||
|
||||
runner._run_scenario(fakes.FakeScenario,
|
||||
"something_went_wrong", self.context, self.args)
|
||||
|
@ -377,9 +377,9 @@ class OSClientsTestCase(test.TestCase):
|
||||
)
|
||||
fake_mistral_url = self.service_catalog.url_for.return_value
|
||||
mock_mistral.client.client.assert_called_once_with(
|
||||
mistral_url=fake_mistral_url,
|
||||
service_type="workflowv2",
|
||||
auth_token=self.fake_keystone.auth_token
|
||||
mistral_url=fake_mistral_url,
|
||||
service_type="workflowv2",
|
||||
auth_token=self.fake_keystone.auth_token
|
||||
)
|
||||
self.assertEqual(fake_mistral, self.clients.cache["mistral"])
|
||||
|
||||
@ -440,7 +440,7 @@ class OSClientsTestCase(test.TestCase):
|
||||
consts.ServiceType.COMPUTE: {},
|
||||
"unknown_service": {}}
|
||||
mock_keystone.return_value = mock.Mock(service_catalog=mock.Mock(
|
||||
get_endpoints=lambda: available_services))
|
||||
get_endpoints=lambda: available_services))
|
||||
clients = osclients.Clients(self.endpoint)
|
||||
|
||||
self.assertEqual(
|
||||
|
Loading…
x
Reference in New Issue
Block a user