[ci] Rework launching cli job
There are 2 critical blocking issues related to cli job: * bug related to db conflict between 2 tests (samples and lib check). We already spent ~ a month reserching the root of an issue and trying to fix it. Unfortunately, we did not succeded. Since we are planning to rework samples mechanism, the applied solution in this patch should not look like a dirty workaround - split launching main cli tests and samples check. To do this, we need to modify out pytest launcher a bit * Rally-CI has two cli jobs for mysql and postgres backends. While porting them to use `tox -ecli` command, passing the proper env variable was missed and sqlite is used now in both cases. Passing the right environment variable is not only required step, setting serial mode of pytest is important too, since we do not want to have conflicts in tests connecting to the one db. Also, this patch removes unused option (--timeout) of pytest launcher. Change-Id: I7fc109ee5785f426211fbff7eb9b4553a3671ba7
This commit is contained in:
parent
2f45502170
commit
b1c84f1058
0
tests/check_samples/__init__.py
Normal file
0
tests/check_samples/__init__.py
Normal file
157
tests/check_samples/test_task_samples.py
Normal file
157
tests/check_samples/test_task_samples.py
Normal file
@ -0,0 +1,157 @@
|
||||
# Copyright 2014: Mirantis Inc.
|
||||
# Copyright 2014: Catalyst IT Ltd.
|
||||
# All Rights Reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import copy
|
||||
import json
|
||||
import mock
|
||||
import os
|
||||
import re
|
||||
import traceback
|
||||
import unittest
|
||||
|
||||
import six
|
||||
|
||||
import rally as rally_module
|
||||
from rally import api
|
||||
from rally.common import broker
|
||||
from rally.common import yamlutils as yaml
|
||||
from rally import plugins
|
||||
from rally.plugins.openstack.context.keystone import users
|
||||
from tests.functional import utils
|
||||
|
||||
|
||||
class TestTaskSamples(unittest.TestCase):
|
||||
|
||||
NUMBER_OF_THREADS = 20
|
||||
|
||||
def _skip(self, validation_output):
|
||||
"""Help to decide do we want to skip this result or not.
|
||||
|
||||
:param validation_output: string representation of the
|
||||
error that we want to check
|
||||
:return: True if we want to skip this error
|
||||
of task sample validation, otherwise False.
|
||||
"""
|
||||
|
||||
skip_lst = ["[Ss]ervice is not available",
|
||||
"is not installed. To install it run",
|
||||
"extension.* is not configured"]
|
||||
for check_str in skip_lst:
|
||||
if re.search(check_str, validation_output) is not None:
|
||||
return True
|
||||
return False
|
||||
|
||||
@plugins.ensure_plugins_are_loaded
|
||||
def test_task_samples_are_valid(self):
|
||||
rally = utils.Rally(force_new_db=True)
|
||||
|
||||
# let's use pre-created users to make TestTaskSamples quicker
|
||||
rapi = api.API(config_file=rally.config_filename)
|
||||
deployment = rapi.deployment._get("MAIN")
|
||||
admin_cred = deployment.get_credentials_for("openstack")["admin"]
|
||||
|
||||
ctx = {"admin": {"credential": admin_cred},
|
||||
"task": {"uuid": self.__class__.__name__,
|
||||
"deployment_uuid": deployment["uuid"]}}
|
||||
user_ctx = users.UserGenerator(ctx)
|
||||
user_ctx.setup()
|
||||
self.addCleanup(user_ctx.cleanup)
|
||||
|
||||
config = deployment["config"]
|
||||
os_creds = config["creds"]["openstack"]
|
||||
|
||||
user = copy.copy(os_creds["admin"])
|
||||
user["username"] = ctx["users"][0]["credential"].username
|
||||
user["password"] = ctx["users"][0]["credential"].password
|
||||
if "project_name" in os_creds["admin"]:
|
||||
# it is Keystone
|
||||
user["project_name"] = ctx["users"][0]["credential"].tenant_name
|
||||
else:
|
||||
user["tenant_name"] = ctx["users"][0]["credential"].tenant_name
|
||||
config["creds"]["openstack"]["users"] = [user]
|
||||
|
||||
rally("deployment destroy MAIN", write_report=False)
|
||||
deployment_cfg = os.path.join(rally.tmp_dir, "new_deployment.json")
|
||||
with open(deployment_cfg, "w") as f:
|
||||
f.write(json.dumps(config))
|
||||
rally("deployment create --name MAIN --filename %s" % deployment_cfg,
|
||||
write_report=False)
|
||||
|
||||
# NOTE(andreykurilin): mock building credential to share one cache of
|
||||
# clients(it will allow to avoid hundreds of redundant
|
||||
# authentications) between validations of different samples
|
||||
deployment = rapi.deployment._get("MAIN")
|
||||
original_get_credentials_for = deployment.get_credentials_for
|
||||
creds_cache = {}
|
||||
|
||||
def get_credentials_for(namespace):
|
||||
if namespace not in creds_cache:
|
||||
creds_cache[namespace] = original_get_credentials_for(
|
||||
namespace)
|
||||
return creds_cache[namespace]
|
||||
|
||||
deployment.get_credentials_for = get_credentials_for
|
||||
|
||||
deployment_patcher = mock.patch("rally.api.objects.Deployment.get")
|
||||
m_deployment = deployment_patcher.start()
|
||||
m_deployment.return_value = deployment
|
||||
self.addCleanup(deployment_patcher.stop)
|
||||
|
||||
# store all failures and print them at once
|
||||
failed_samples = {}
|
||||
|
||||
def publisher(queue):
|
||||
"""List all samples and render task configs"""
|
||||
samples_path = os.path.join(
|
||||
os.path.dirname(rally_module.__file__), os.pardir,
|
||||
"samples", "tasks")
|
||||
|
||||
for dirname, dirnames, filenames in os.walk(samples_path):
|
||||
# NOTE(rvasilets): Skip by suggest of boris-42 because in
|
||||
# future we don't what to maintain this dir
|
||||
if dirname.find("tempest-do-not-run-against-production") != -1:
|
||||
continue
|
||||
for filename in filenames:
|
||||
full_path = os.path.join(dirname, filename)
|
||||
|
||||
# NOTE(hughsaunders): Skip non config files
|
||||
# (bug https://bugs.launchpad.net/rally/+bug/1314369)
|
||||
if os.path.splitext(filename)[1] != ".json":
|
||||
continue
|
||||
with open(full_path) as task_file:
|
||||
input_task = task_file.read()
|
||||
rendered_task = rapi.task.render_template(
|
||||
task_template=input_task)
|
||||
queue.append((full_path, rendered_task))
|
||||
|
||||
def consumer(_cache, sample):
|
||||
"""Validate one sample"""
|
||||
full_path, rendered_task = sample
|
||||
task_config = yaml.safe_load(rendered_task)
|
||||
try:
|
||||
rapi.task.validate(deployment="MAIN",
|
||||
config=task_config)
|
||||
except Exception as e:
|
||||
if not self._skip(six.text_type(e)):
|
||||
failed_samples[full_path] = traceback.format_exc()
|
||||
|
||||
broker.run(publisher, consumer, self.NUMBER_OF_THREADS)
|
||||
|
||||
if failed_samples:
|
||||
self.fail("Validation failed on the one or several samples. "
|
||||
"See details below:\n%s" %
|
||||
"".join(["\n======\n%s\n\n%s\n" % (k, v)
|
||||
for k, v in failed_samples.items()]))
|
@ -25,7 +25,7 @@ PYTEST_ARGUMENTS = ("py.test" # base command
|
||||
" --html=%(html_report)s" # html report
|
||||
" --self-contained-html" # embedded css
|
||||
" --durations=10" # get a list of the slowest 10 tests
|
||||
" -n auto" # launch tests in parallel
|
||||
" -n %(concurrency)s" # launch tests in parallel
|
||||
" %(path)s"
|
||||
)
|
||||
|
||||
@ -42,9 +42,8 @@ def main(args):
|
||||
parser.add_argument("--posargs", metavar="<str>", type=str, default="",
|
||||
help="TOX posargs. Currently supported only string to "
|
||||
"partial test or tests group to launch.")
|
||||
parser.add_argument("--timeout", metavar="<seconds>", type=int, default=60,
|
||||
help="Timeout for individual test execution. "
|
||||
"Defaults to 60")
|
||||
parser.add_argument("--concurrency", metavar="<N>", type=int,
|
||||
help="Number of parallel processes.")
|
||||
args = parser.parse_args(args[1:])
|
||||
|
||||
# We allow only one parameter - path to partial test or tests group
|
||||
@ -99,7 +98,7 @@ def main(args):
|
||||
|
||||
args = PYTEST_ARGUMENTS % {"html_report": pytest_report,
|
||||
"path": path,
|
||||
"timeout": args.timeout}
|
||||
"concurrency": args.concurrency or "auto"}
|
||||
try:
|
||||
subprocess.check_call(args.split(" "),
|
||||
stderr=subprocess.STDOUT)
|
||||
|
2
tox.ini
2
tox.ini
@ -57,7 +57,7 @@ commands = oslo_debug_helper -t tests {posargs}
|
||||
sitepackages = True
|
||||
commands =
|
||||
find . -type f -name "*.pyc" -delete
|
||||
python {toxinidir}/tests/ci/pytest_launcher.py "tests/functional" --timeout -1 --posargs={posargs}
|
||||
{toxinidir}/tests/ci/rally_cli_job.sh
|
||||
|
||||
[testenv:cover]
|
||||
commands = {toxinidir}/tests/ci/cover.sh {posargs}
|
||||
|
Loading…
Reference in New Issue
Block a user