Metadata and restructure
Moving some methods into Tools to cleanup browbeat.py + Adding Ansible task to test regather Change-Id: I4a562ae77a90528df2e7f8779895b1627c9dff11
This commit is contained in:
parent
27c234c8e4
commit
c9b8d88697
@ -5,3 +5,12 @@
|
||||
"source {{ ansible_env.HOME }}/browbeat-venv/bin/activate; \
|
||||
cd {{ ansible_env.HOME }}/browbeat/; \
|
||||
python browbeat.py rally > {{ ansible_env.HOME }}/browbeat/results/browbeat_run.log"
|
||||
|
||||
- name: Run Browbeat Test Regather
|
||||
shell:
|
||||
"source {{ ansible_env.HOME }}/browbeat-venv/bin/activate; \
|
||||
cd {{ ansible_env.HOME }}/browbeat/; \
|
||||
rm metadata/software-metadata.json; \
|
||||
rm metadata/hardware-metadata.json; \
|
||||
rm metadata/environment-metadata.json; \
|
||||
python browbeat.py rally > {{ ansible_env.HOME }}/browbeat/results/browbeat_run.log"
|
||||
|
@ -6,6 +6,7 @@ elasticsearch:
|
||||
enabled: {{ elastic_enabled_template }}
|
||||
host: {{ elastic_host_template }}
|
||||
port: 9200
|
||||
regather: true
|
||||
metadata_files:
|
||||
- name: hardware-metadata
|
||||
file: metadata/hardware-metadata.json
|
||||
|
@ -8,6 +8,7 @@ elasticsearch:
|
||||
enabled: {{ elastic_enabled_template }}
|
||||
host: {{ elastic_host_template }}
|
||||
port: 9200
|
||||
regather: true
|
||||
metadata_files:
|
||||
- name: hardware-metadata
|
||||
file: metadata/hardware-metadata.json
|
||||
|
@ -4,9 +4,10 @@ browbeat:
|
||||
rerun: 1
|
||||
cloud_name: openstack
|
||||
elasticsearch:
|
||||
enabled: false
|
||||
enabled: true
|
||||
host: 1.1.1.1
|
||||
port: 9200
|
||||
regather: true
|
||||
metadata_files:
|
||||
- name: hardware-metadata
|
||||
file: metadata/hardware-metadata.json
|
||||
@ -39,7 +40,7 @@ grafana:
|
||||
enabled: false
|
||||
snapshot_compute: false
|
||||
perfkit:
|
||||
enabled: true
|
||||
enabled: false
|
||||
sleep_before: 0
|
||||
sleep_after: 0
|
||||
venv: /home/stack/perfkit-venv/bin/activate
|
||||
@ -163,7 +164,7 @@ rally:
|
||||
times: 100
|
||||
scenarios:
|
||||
- name: create-list-network
|
||||
enabled: false
|
||||
enabled: true
|
||||
file: rally/neutron/neutron-create-list-network-cc.yml
|
||||
sla_max_seconds: 30
|
||||
sla_max_failure: 0
|
||||
|
73
browbeat.py
73
browbeat.py
@ -20,65 +20,16 @@ import lib.Tools
|
||||
import argparse
|
||||
import logging
|
||||
import sys
|
||||
import yaml
|
||||
import time
|
||||
import datetime
|
||||
import os
|
||||
from pykwalify import core as pykwalify_core
|
||||
from pykwalify import errors as pykwalify_errors
|
||||
|
||||
_workload_opts = ['perfkit', 'rally', 'shaker']
|
||||
_config_file = 'browbeat-config.yaml'
|
||||
debug_log_file = 'log/debug.log'
|
||||
|
||||
def _load_config(path, _logger):
|
||||
try:
|
||||
stream = open(path, 'r')
|
||||
except IOError:
|
||||
_logger.error("Configuration file {} passed is missing".format(path))
|
||||
exit(1)
|
||||
config = yaml.load(stream)
|
||||
stream.close()
|
||||
validate_yaml(config, _logger)
|
||||
return config
|
||||
|
||||
def validate_yaml(config, _logger):
|
||||
_logger.info("Validating the configuration file passed by the user")
|
||||
stream = open("lib/validate.yaml", 'r')
|
||||
schema = yaml.load(stream)
|
||||
check = pykwalify_core.Core(source_data=config, schema_data=schema)
|
||||
try:
|
||||
check.validate(raise_exception=True)
|
||||
_logger.info("Validation successful")
|
||||
except pykwalify_errors.SchemaError as e:
|
||||
_logger.error("Schema Validation failed")
|
||||
raise Exception('File does not conform to schema: {}'.format(e))
|
||||
|
||||
def _run_workload_provider(provider, config):
|
||||
_logger = logging.getLogger('browbeat')
|
||||
if provider == "perfkit":
|
||||
perfkit = lib.PerfKit.PerfKit(config)
|
||||
perfkit.start_workloads()
|
||||
elif provider == "rally":
|
||||
rally = lib.Rally.Rally(config)
|
||||
rally.start_workloads()
|
||||
elif provider == "shaker":
|
||||
shaker = lib.Shaker.Shaker(config)
|
||||
shaker.run_shaker()
|
||||
else:
|
||||
_logger.error("Unknown workload provider: {}".format(provider))
|
||||
|
||||
def check_metadata(config, _logger):
|
||||
_logger.debug("Checking if configured metadata files are present")
|
||||
meta = config['elasticsearch']['metadata_files']
|
||||
for _meta in meta:
|
||||
if not os.path.isfile(_meta['file']):
|
||||
_logger.error("Metadata file {} is not present".format(_meta['file']))
|
||||
return False
|
||||
return True
|
||||
|
||||
|
||||
def main():
|
||||
tools = lib.Tools.Tools()
|
||||
parser = argparse.ArgumentParser(
|
||||
description="Browbeat Performance and Scale testing for Openstack")
|
||||
parser.add_argument(
|
||||
@ -111,7 +62,7 @@ def main():
|
||||
_logger.debug("CLI Args: {}".format(_cli_args))
|
||||
|
||||
# Load Browbeat yaml config file:
|
||||
_config = _load_config(_cli_args.setup, _logger)
|
||||
_config = tools._load_config(_cli_args.setup)
|
||||
|
||||
# Default to all workloads
|
||||
if _cli_args.workloads == []:
|
||||
@ -131,27 +82,21 @@ def main():
|
||||
_logger.info("Browbeat UUID: {}".format(browbeat_uuid))
|
||||
if _config['elasticsearch']['enabled']:
|
||||
_logger.info("Checking for Metadata")
|
||||
metadata_exists = check_metadata(_config, _logger)
|
||||
metadata_exists = tools.check_metadata()
|
||||
if not metadata_exists:
|
||||
_logger.error("Elasticsearch has been enabled but"
|
||||
" metadata files do not exist")
|
||||
_logger.info("Gathering Metadata")
|
||||
os.putenv("ANSIBLE_SSH_ARGS"," -F {}".format(_config['ansible']['ssh_config']))
|
||||
tools = lib.Tools.Tools(_config)
|
||||
ansible_cmd = \
|
||||
'ansible-playbook -i {} {}' \
|
||||
.format(_config['ansible']['hosts'], _config['ansible']['metadata'])
|
||||
tools.run_cmd(ansible_cmd)
|
||||
if not check_metadata(_config, _logger):
|
||||
_logger.warning("Metadata could not be gathered")
|
||||
exit(1)
|
||||
else:
|
||||
_logger.info("Metadata about cloud has been gathered")
|
||||
tools.gather_metadata()
|
||||
elif _config['elasticsearch']['regather'] :
|
||||
_logger.info("Regathering Metadata")
|
||||
tools.gather_metadata()
|
||||
|
||||
_logger.info("Running workload(s): {}".format(','.join(_cli_args.workloads)))
|
||||
for wkld_provider in _cli_args.workloads:
|
||||
if wkld_provider in _config:
|
||||
if _config[wkld_provider]['enabled']:
|
||||
_run_workload_provider(wkld_provider, _config)
|
||||
tools._run_workload_provider(wkld_provider)
|
||||
else:
|
||||
_logger.warning("{} is not enabled in {}".format(wkld_provider,
|
||||
_cli_args.setup))
|
||||
|
@ -8,6 +8,7 @@ elasticsearch:
|
||||
enabled: false
|
||||
host: 1.1.1.1
|
||||
port: 9200
|
||||
regather: true
|
||||
metadata_files:
|
||||
- name: hardware-metadata
|
||||
file: metadata/hardware-metadata.json
|
||||
|
66
lib/Tools.py
66
lib/Tools.py
@ -10,10 +10,15 @@
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
import PerfKit
|
||||
import Rally
|
||||
import Shaker
|
||||
import logging
|
||||
import os
|
||||
import subprocess
|
||||
|
||||
import yaml
|
||||
from pykwalify import core as pykwalify_core
|
||||
from pykwalify import errors as pykwalify_errors
|
||||
|
||||
class Tools(object):
|
||||
|
||||
@ -62,3 +67,62 @@ class Tools(object):
|
||||
self.logger.error("Error creating the results directory: {}".format(err))
|
||||
return False
|
||||
return the_directory
|
||||
|
||||
def _load_config(self, path):
|
||||
try:
|
||||
stream = open(path, 'r')
|
||||
except IOError:
|
||||
self.logger.error("Configuration file {} passed is missing".format(path))
|
||||
exit(1)
|
||||
config = yaml.load(stream)
|
||||
stream.close()
|
||||
self.config = config
|
||||
self.validate_yaml()
|
||||
return config
|
||||
|
||||
def validate_yaml(self):
|
||||
self.logger.info("Validating the configuration file passed by the user")
|
||||
stream = open("lib/validate.yaml", 'r')
|
||||
schema = yaml.load(stream)
|
||||
check = pykwalify_core.Core(source_data=self.config, schema_data=schema)
|
||||
try:
|
||||
check.validate(raise_exception=True)
|
||||
self.logger.info("Validation successful")
|
||||
except pykwalify_errors.SchemaError as e:
|
||||
self.logger.error("Schema Validation failed")
|
||||
raise Exception('File does not conform to schema: {}'.format(e))
|
||||
|
||||
def _run_workload_provider(self, provider):
|
||||
self.logger = logging.getLogger('browbeat')
|
||||
if provider == "perfkit":
|
||||
perfkit = PerfKit.PerfKit(self.config)
|
||||
perfkit.start_workloads()
|
||||
elif provider == "rally":
|
||||
rally = Rally.Rally(self.config)
|
||||
rally.start_workloads()
|
||||
elif provider == "shaker":
|
||||
shaker = Shaker.Shaker(self.config)
|
||||
shaker.run_shaker()
|
||||
else:
|
||||
self.logger.error("Unknown workload provider: {}".format(provider))
|
||||
|
||||
def check_metadata(self):
|
||||
meta = self.config['elasticsearch']['metadata_files']
|
||||
for _meta in meta:
|
||||
if not os.path.isfile(_meta['file']):
|
||||
self.logger.error("Metadata file {} is not present".format(_meta['file']))
|
||||
return False
|
||||
return True
|
||||
|
||||
def gather_metadata(self):
|
||||
os.putenv("ANSIBLE_SSH_ARGS"," -F {}".format(self.config['ansible']['ssh_config']))
|
||||
ansible_cmd = \
|
||||
'ansible-playbook -i {} {}' \
|
||||
.format(self.config['ansible']['hosts'], self.config['ansible']['metadata'])
|
||||
self.run_cmd(ansible_cmd)
|
||||
if not self.check_metadata():
|
||||
self.logger.warning("Metadata could not be gathered")
|
||||
return False
|
||||
else:
|
||||
self.logger.info("Metadata about cloud has been gathered")
|
||||
return True
|
||||
|
@ -27,6 +27,9 @@ mapping:
|
||||
type: str
|
||||
required: True
|
||||
pattern: ^([a-zA-Z0-9]|[a-zA-Z0-9][a-zA-Z0-9\-]{0,61}[a-zA-Z0-9])(\.([a-zA-Z0-9]|[a-zA-Z0-9][a-zA-Z0-9\-]{0,61}[a-zA-Z0-9]))*$
|
||||
regather:
|
||||
type: bool
|
||||
required: True
|
||||
port:
|
||||
type: int
|
||||
required: True
|
||||
|
Loading…
Reference in New Issue
Block a user