Remove option all in dynamic workloads

Dynamic workloads has an option all in browbeat-config.yaml,
which doesn't really makes sense because it runs only VM and trunk
dynamic workloads. This patch replaces all with an option called
all_vm_and_trunk, for better readability.

Change-Id: I9e3a73a9fb98fe46c830ad56b6a3cdd4a8c2b863
This commit is contained in:
Sanjay Chari 2022-11-09 11:44:44 +05:30
parent 8903bb84a4
commit b20acab937
2 changed files with 22 additions and 16 deletions

View File

@ -709,23 +709,23 @@ workloads:
# created as part of rally context for dynamic workloads. These external
# networks will be used in a round robin fashion by the iterations.
num_external_networks: 16
# workloads can be 'all', a single workload(Eg. : create_delete_servers),
# workloads can contain a single workload(Eg. : create_delete_servers),
# or a comma separated string(Eg. : create_delete_servers,migrate_servers).
# Currently supported workloads : create_delete_servers, migrate_servers
# swap_floating_ips_between_servers, stop_start_servers, create_loadbalancers,
# delete_loadbalancers, delete_members_random_lb, pod_fip_simulation,
# add_subports_to_random_trunks, delete_subports_from_random_trunks,
# swap_floating_ips_between_random_subports, provider_netcreate_nova_boot_ping,
# provider_net_nova_boot_ping, provider_net_nova_delete, e2e_kube_burner, ocp_on_osp
# Note: Octavia, Provider, e2e_kube_burner and ocp_on_osp scenarios are not included in 'all'
# by default, and have to be included separately.
# swap_floating_ips_between_random_subports,
# all_vm_and_trunk(will run all vm and trunk dynamic workloads),
# provider_netcreate_nova_boot_ping, provider_net_nova_boot_ping, provider_net_nova_delete,
# e2e_kube_burner, ocp_on_osp
# Steps for running ocp_on_osp workload, given that underlying openstack has been
# deployed with necessary dependencies:
# 1) Pass your pull secret in ocp_on_osp/vars/shift_stack_vars.yaml
# 2) If you want to change any default params(like ocp_version, worker count, master and worker flavors, etc), refer to ocp_on_osp/vars/shift_stack_vars.yml
# 3) Run "ansible-playbook -vvv ocp_on_osp/ocp_bootstrap.yml"
# 4) And then run ocp_on_osp workload. NOTE: use "ansible-playbook -vvv ocp_on_osp/ocp_cleanup.yml" for the clean up.
workloads: all
workloads: all_vm_and_trunk
file: rally/rally-plugins/dynamic-workloads/dynamic_workload.yml
- name: dynamic-workloads-min

View File

@ -79,7 +79,9 @@ class DynamicWorkload(vm.VMDynamicScenario, trunk.TrunkDynamicScenario,
self.security_group = self.create_sec_group_with_icmp_ssh()
self.log_info("security group {} created for this iteration".format(self.security_group))
if(workloads == "all" or "migrate_servers" in workloads_list or
run_all_vm_and_trunk_workloads = "all_vm_and_trunk" in workloads_list
if(run_all_vm_and_trunk_workloads or "migrate_servers" in workloads_list or
"swap_floating_ips_between_servers" in workloads_list or
"stop_start_servers" in workloads_list or
"pod_fip_simulation" in workloads_list or
@ -103,38 +105,42 @@ class DynamicWorkload(vm.VMDynamicScenario, trunk.TrunkDynamicScenario,
except AttributeError:
DynamicWorkload.browbeat_dir = os.getcwd()
if workloads == "all" or "create_delete_servers" in workloads_list:
if run_all_vm_and_trunk_workloads or "create_delete_servers" in workloads_list:
self.boot_servers(smallest_image, smallest_flavor, num_create_vms,
subnet_create_args=subnet_create_args)
self.delete_random_servers(num_delete_vms)
if(workloads == "all" or "migrate_servers" in workloads_list or
if(run_all_vm_and_trunk_workloads or "migrate_servers" in workloads_list or
"swap_floating_ips_between_servers" in workloads_list or
"stop_start_servers" in workloads_list):
self.boot_servers_with_fip(smallest_image, smallest_flavor, context_ext_net_id,
num_vms_to_create_with_fip,
network_create_args, subnet_create_args, **kwargs)
if workloads == "all" or "migrate_servers" in workloads_list:
if run_all_vm_and_trunk_workloads or "migrate_servers" in workloads_list:
self.migrate_servers_with_fip(num_vms_to_migrate)
if workloads == "all" or "swap_floating_ips_between_servers" in workloads_list:
if(run_all_vm_and_trunk_workloads or
"swap_floating_ips_between_servers" in workloads_list):
self.swap_floating_ips_between_servers()
if workloads == "all" or "stop_start_servers" in workloads_list:
if run_all_vm_and_trunk_workloads or "stop_start_servers" in workloads_list:
self.stop_start_servers_with_fip(num_stop_start_vms)
if workloads == "all" or "pod_fip_simulation" in workloads_list:
if run_all_vm_and_trunk_workloads or "pod_fip_simulation" in workloads_list:
self.pod_fip_simulation(context_ext_net_id, trunk_image, trunk_flavor, smallest_image,
smallest_flavor, num_initial_subports, num_trunk_vms)
if workloads == "all" or "add_subports_to_random_trunks" in workloads_list:
if(run_all_vm_and_trunk_workloads or
"add_subports_to_random_trunks" in workloads_list):
self.add_subports_to_random_trunks(num_add_subports_trunks, num_add_subports)
if workloads == "all" or "delete_subports_from_random_trunks" in workloads_list:
if(run_all_vm_and_trunk_workloads or
"delete_subports_from_random_trunks" in workloads_list):
self.delete_subports_from_random_trunks(num_delete_subports_trunks, num_delete_subports)
if workloads == "all" or "swap_floating_ips_between_random_subports" in workloads_list:
if(run_all_vm_and_trunk_workloads or
"swap_floating_ips_between_random_subports" in workloads_list):
self.swap_floating_ips_between_random_subports()
if "create_loadbalancers" in workloads_list: