Files
rally/tests/unit/common/db/test_migrations.py
Andrey Kurilin 7165f31726 Port inner stuff to the new task format
* use new format of a runner section
  the schemas of existing plugins are changed to not include "type"

* use "contexts" key instead of "context"
  note: the database model still operates the word "context". hope it
        will be fixed soon while extendind abilities of contexts

* use new format of a hook section

Change-Id: I2ef6ba7a24b542fb001bce378cadf8c83c774b01
2017-10-06 15:54:41 +03:00

2100 lines
88 KiB
Python

# Copyright (c) 2016 Mirantis Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
"""Tests for DB migration."""
import copy
import json
import pickle
import pprint
import uuid
import alembic
import mock
from oslo_db.sqlalchemy import test_migrations
from oslo_db.sqlalchemy import utils as db_utils
from oslo_utils import timeutils
import six
import sqlalchemy as sa
import rally
from rally.common import db
from rally.common.db.sqlalchemy import api
from rally.common.db.sqlalchemy import models
from rally import consts
from rally.deployment.engines import existing
from tests.unit.common.db import test_migrations_base
from tests.unit import test as rtest
class MigrationTestCase(rtest.DBTestCase,
test_migrations.ModelsMigrationsSync):
"""Test for checking of equality models state and migrations.
For the opportunistic testing you need to set up a db named
'openstack_citest' with user 'openstack_citest' and password
'openstack_citest' on localhost.
The test will then use that db and user/password combo to run the tests.
For PostgreSQL on Ubuntu this can be done with the following commands::
sudo -u postgres psql
postgres=# create user openstack_citest with createdb login password
'openstack_citest';
postgres=# create database openstack_citest with owner
openstack_citest;
For MySQL on Ubuntu this can be done with the following commands::
mysql -u root
>create database openstack_citest;
>grant all privileges on openstack_citest.* to
openstack_citest@localhost identified by 'openstack_citest';
Output is a list that contains information about differences between db and
models. Output example::
[('add_table',
Table('bat', MetaData(bind=None),
Column('info', String(), table=<bat>), schema=None)),
('remove_table',
Table(u'bar', MetaData(bind=None),
Column(u'data', VARCHAR(), table=<bar>), schema=None)),
('add_column',
None,
'foo',
Column('data', Integer(), table=<foo>)),
('remove_column',
None,
'foo',
Column(u'old_data', VARCHAR(), table=None)),
[('modify_nullable',
None,
'foo',
u'x',
{'existing_server_default': None,
'existing_type': INTEGER()},
True,
False)]]
* ``remove_*`` means that there is extra table/column/constraint in db;
* ``add_*`` means that it is missing in db;
* ``modify_*`` means that on column in db is set wrong
type/nullable/server_default. Element contains information:
- what should be modified,
- schema,
- table,
- column,
- existing correct column parameters,
- right value,
- wrong value.
"""
def setUp(self):
# we change DB metadata in tests so we reload
# models to refresh the metadata to it's original state
six.moves.reload_module(rally.common.db.sqlalchemy.models)
super(MigrationTestCase, self).setUp()
self.alembic_config = api._alembic_config()
self.engine = api.get_engine()
# remove everything from DB and stamp it as 'base'
# so that migration (i.e. upgrade up to 'head')
# will actually take place
db.schema_cleanup()
db.schema_stamp("base")
def db_sync(self, engine):
db.schema_upgrade()
def get_engine(self):
return self.engine
def get_metadata(self):
return models.BASE.metadata
def include_object(self, object_, name, type_, reflected, compare_to):
if type_ == "table" and name == "alembic_version":
return False
return super(MigrationTestCase, self).include_object(
object_, name, type_, reflected, compare_to)
def _create_fake_model(self, table_name):
type(
"FakeModel",
(models.BASE, models.RallyBase),
{"__tablename__": table_name,
"id": sa.Column(sa.Integer, primary_key=True,
autoincrement=True)}
)
def _get_metadata_diff(self):
with self.get_engine().connect() as conn:
opts = {
"include_object": self.include_object,
"compare_type": self.compare_type,
"compare_server_default": self.compare_server_default,
}
mc = alembic.migration.MigrationContext.configure(conn, opts=opts)
# compare schemas and fail with diff, if it"s not empty
diff = self.filter_metadata_diff(
alembic.autogenerate.compare_metadata(mc, self.get_metadata()))
return diff
@mock.patch("rally.common.db.sqlalchemy.api.Connection.schema_stamp")
def test_models_sync(self, mock_connection_schema_stamp):
# drop all tables after a test run
self.addCleanup(db.schema_cleanup)
# run migration scripts
self.db_sync(self.get_engine())
diff = self._get_metadata_diff()
if diff:
msg = pprint.pformat(diff, indent=2, width=20)
self.fail(
"Models and migration scripts aren't in sync:\n%s" % msg)
@mock.patch("rally.common.db.sqlalchemy.api.Connection.schema_stamp")
def test_models_sync_negative__missing_table_in_script(
self, mock_connection_schema_stamp):
# drop all tables after a test run
self.addCleanup(db.schema_cleanup)
self._create_fake_model("fake_model")
# run migration scripts
self.db_sync(self.get_engine())
diff = self._get_metadata_diff()
self.assertEqual(1, len(diff))
action, object = diff[0]
self.assertEqual("add_table", action)
self.assertIsInstance(object, sa.Table)
self.assertEqual("fake_model", object.name)
@mock.patch("rally.common.db.sqlalchemy.api.Connection.schema_stamp")
def test_models_sync_negative__missing_model_in_metadata(
self, mock_connection_schema_stamp):
# drop all tables after a test run
self.addCleanup(db.schema_cleanup)
table = self.get_metadata().tables["workers"]
self.get_metadata().remove(table)
# run migration scripts
self.db_sync(self.get_engine())
diff = self._get_metadata_diff()
self.assertEqual(1, len(diff))
action, object = diff[0]
self.assertEqual("remove_table", action)
self.assertIsInstance(object, sa.Table)
self.assertEqual("workers", object.name)
class MigrationWalkTestCase(rtest.DBTestCase,
test_migrations_base.BaseWalkMigrationMixin):
"""Test case covers upgrade method in migrations."""
def setUp(self):
super(MigrationWalkTestCase, self).setUp()
self.engine = api.get_engine()
def assertColumnExists(self, engine, table, column):
t = db_utils.get_table(engine, table)
self.assertIn(column, t.c)
def assertColumnsExists(self, engine, table, columns):
for column in columns:
self.assertColumnExists(engine, table, column)
def assertColumnCount(self, engine, table, columns):
t = db_utils.get_table(engine, table)
self.assertEqual(len(columns), len(t.columns))
def assertColumnNotExists(self, engine, table, column):
t = db_utils.get_table(engine, table)
self.assertNotIn(column, t.c)
def assertIndexExists(self, engine, table, index):
t = db_utils.get_table(engine, table)
index_names = [idx.name for idx in t.indexes]
self.assertIn(index, index_names)
def assertColumnType(self, engine, table, column, sqltype):
t = db_utils.get_table(engine, table)
col = getattr(t.c, column)
self.assertIsInstance(col.type, sqltype)
def assertIndexMembers(self, engine, table, index, members):
self.assertIndexExists(engine, table, index)
t = db_utils.get_table(engine, table)
index_columns = None
for idx in t.indexes:
if idx.name == index:
index_columns = idx.columns.keys()
break
self.assertEqual(sorted(members), sorted(index_columns))
def test_walk_versions(self):
self.walk_versions(self.engine)
def _check_3177d36ea270(self, engine, data):
self.assertEqual(
"3177d36ea270", api.get_backend().schema_revision(engine=engine))
self.assertColumnExists(engine, "deployments", "credentials")
self.assertColumnNotExists(engine, "deployments", "admin")
self.assertColumnNotExists(engine, "deployments", "users")
def _pre_upgrade_54e844ebfbc3(self, engine):
self._54e844ebfbc3_deployments = {
# right config which should not be changed after migration
"should-not-be-changed-1": {
"admin": {"username": "admin",
"password": "passwd",
"project_name": "admin"},
"auth_url": "http://example.com:5000/v3",
"region_name": "RegionOne",
"type": "ExistingCloud"},
# right config which should not be changed after migration
"should-not-be-changed-2": {
"admin": {"username": "admin",
"password": "passwd",
"tenant_name": "admin"},
"users": [{"username": "admin",
"password": "passwd",
"tenant_name": "admin"}],
"auth_url": "http://example.com:5000/v2.0",
"region_name": "RegionOne",
"type": "ExistingCloud"},
# not ExistingCloud config which should not be changed
"should-not-be-changed-3": {
"url": "example.com",
"type": "Something"},
# normal config created with "fromenv" feature
"from-env": {
"admin": {"username": "admin",
"password": "passwd",
"tenant_name": "admin",
"project_domain_name": "",
"user_domain_name": ""},
"auth_url": "http://example.com:5000/v2.0",
"region_name": "RegionOne",
"type": "ExistingCloud"},
# public endpoint + keystone v3 config with tenant_name
"ksv3_public": {
"admin": {"username": "admin",
"password": "passwd",
"tenant_name": "admin",
"user_domain_name": "bla",
"project_domain_name": "foo"},
"auth_url": "http://example.com:5000/v3",
"region_name": "RegionOne",
"type": "ExistingCloud",
"endpoint_type": "public"},
# internal endpoint + existing_users
"existing_internal": {
"admin": {"username": "admin",
"password": "passwd",
"tenant_name": "admin"},
"users": [{"username": "admin",
"password": "passwd",
"tenant_name": "admin",
"project_domain_name": "",
"user_domain_name": ""}],
"auth_url": "http://example.com:5000/v2.0",
"region_name": "RegionOne",
"type": "ExistingCloud",
"endpoint_type": "internal"},
}
deployment_table = db_utils.get_table(engine, "deployments")
deployment_status = consts.DeployStatus.DEPLOY_FINISHED
with engine.connect() as conn:
for deployment in self._54e844ebfbc3_deployments:
conf = json.dumps(self._54e844ebfbc3_deployments[deployment])
conn.execute(
deployment_table.insert(),
[{"uuid": deployment, "name": deployment,
"config": conf,
"enum_deployments_status": deployment_status,
"credentials": six.b(json.dumps([])),
"users": six.b(json.dumps([]))
}])
def _check_54e844ebfbc3(self, engine, data):
self.assertEqual("54e844ebfbc3",
api.get_backend().schema_revision(engine=engine))
original_deployments = self._54e844ebfbc3_deployments
deployment_table = db_utils.get_table(engine, "deployments")
with engine.connect() as conn:
deployments_found = conn.execute(
deployment_table.select()).fetchall()
for deployment in deployments_found:
# check deployment
self.assertIn(deployment.uuid, original_deployments)
self.assertIn(deployment.name, original_deployments)
config = json.loads(deployment.config)
if config != original_deployments[deployment.uuid]:
if deployment.uuid.startswith("should-not-be-changed"):
self.fail("Config of deployment '%s' is changes, but "
"should not." % deployment.uuid)
endpoint_type = (original_deployments[
deployment.uuid].get("endpoint_type"))
if endpoint_type in (None, "public"):
self.assertNotIn("endpoint_type", config)
else:
self.assertIn("endpoint_type", config)
self.assertEqual(endpoint_type,
config["endpoint_type"])
existing.ExistingCloud({"config": config}).validate()
else:
if not deployment.uuid.startswith("should-not-be-changed"):
self.fail("Config of deployment '%s' is not changes, "
"but should." % deployment.uuid)
# this deployment created at _pre_upgrade step is not needed
# anymore and we can remove it
conn.execute(
deployment_table.delete().where(
deployment_table.c.uuid == deployment.uuid)
)
def _pre_upgrade_08e1515a576c(self, engine):
self._08e1515a576c_logs = [
{"pre": "No such file name",
"post": {"etype": IOError.__name__, "msg": "No such file name"}},
{"pre": "Task config is invalid: bla",
"post": {"etype": "InvalidTaskException",
"msg": "Task config is invalid: bla"}},
{"pre": "Failed to load task foo",
"post": {"etype": "FailedToLoadTask",
"msg": "Failed to load task foo"}},
{"pre": ["SomeCls", "msg", json.dumps(
["File some1.py, line ...\n",
"File some2.py, line ...\n"])],
"post": {"etype": "SomeCls",
"msg": "msg",
"trace": "Traceback (most recent call last):\n"
"File some1.py, line ...\n"
"File some2.py, line ...\nSomeCls: msg"}},
]
deployment_table = db_utils.get_table(engine, "deployments")
task_table = db_utils.get_table(engine, "tasks")
self._08e1515a576c_deployment_uuid = "08e1515a576c-uuuu-uuuu-iiii-dddd"
with engine.connect() as conn:
conn.execute(
deployment_table.insert(),
[{"uuid": self._08e1515a576c_deployment_uuid,
"name": self._08e1515a576c_deployment_uuid,
"config": six.b("{}"),
"enum_deployments_status":
consts.DeployStatus.DEPLOY_FINISHED,
"credentials": six.b(json.dumps([])),
"users": six.b(json.dumps([]))
}])
for i in range(0, len(self._08e1515a576c_logs)):
log = json.dumps(self._08e1515a576c_logs[i]["pre"])
conn.execute(
task_table.insert(),
[{"uuid": i,
"verification_log": log,
"status": "failed",
"enum_tasks_status": "failed",
"deployment_uuid": self._08e1515a576c_deployment_uuid
}])
def _check_08e1515a576c(self, engine, data):
self.assertEqual("08e1515a576c",
api.get_backend().schema_revision(engine=engine))
tasks = self._08e1515a576c_logs
deployment_table = db_utils.get_table(engine, "deployments")
task_table = db_utils.get_table(engine, "tasks")
with engine.connect() as conn:
tasks_found = conn.execute(task_table.select()).fetchall()
for task in tasks_found:
actual_log = json.loads(task.verification_log)
self.assertIsInstance(actual_log, dict)
expected = tasks[int(task.uuid)]["post"]
for key in expected:
self.assertEqual(expected[key], actual_log[key])
conn.execute(
task_table.delete().where(task_table.c.uuid == task.uuid))
deployment_uuid = self._08e1515a576c_deployment_uuid
conn.execute(
deployment_table.delete().where(
deployment_table.c.uuid == deployment_uuid))
def _pre_upgrade_e654a0648db0(self, engine):
deployment_table = db_utils.get_table(engine, "deployments")
task_table = db_utils.get_table(engine, "tasks")
taskresult_table = db_utils.get_table(engine, "task_results")
self._e654a0648db0_task_uuid = str(uuid.uuid4())
self._e654a0648db0_deployment_uuid = str(uuid.uuid4())
with engine.connect() as conn:
conn.execute(
deployment_table.insert(),
[{
"uuid": self._e654a0648db0_deployment_uuid,
"name": self._e654a0648db0_deployment_uuid,
"config": "{}",
"enum_deployments_status": consts.DeployStatus.DEPLOY_INIT,
"credentials": six.b(json.dumps([])),
"users": six.b(json.dumps([]))
}]
)
conn.execute(
task_table.insert(),
[{
"uuid": self._e654a0648db0_task_uuid,
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"status": consts.TaskStatus.FINISHED,
"verification_log": json.dumps({}),
"tag": "test_tag",
"deployment_uuid": self._e654a0648db0_deployment_uuid
}]
)
conn.execute(
taskresult_table.insert(), [
{
"task_uuid": self._e654a0648db0_task_uuid,
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"key": json.dumps({
"name": "test_scenario",
"pos": 0,
"kw": {
"args": {"a": "A"},
"runner": {"type": "theRunner"},
"context": {"c": "C"},
"sla": {"s": "S"}
}
}),
"data": json.dumps({
"raw": [
{"error": "e", "duration": 3},
{"duration": 1},
{"duration": 8},
],
"load_duration": 42,
"full_duration": 142,
"sla": [{"success": True}, {"success": False}]
})
}
]
)
def _check_e654a0648db0(self, engine, data):
self.assertEqual(
"e654a0648db0", api.get_backend().schema_revision(engine=engine))
task_table = db_utils.get_table(engine, "tasks")
subtask_table = db_utils.get_table(engine, "subtasks")
workload_table = db_utils.get_table(engine, "workloads")
workloaddata_table = db_utils.get_table(engine, "workloaddata")
tag_table = db_utils.get_table(engine, "tags")
deployment_table = db_utils.get_table(engine, "deployments")
with engine.connect() as conn:
# Check task
tasks_found = conn.execute(
task_table.select().
where(task_table.c.uuid == self._e654a0648db0_task_uuid)
).fetchall()
self.assertEqual(1, len(tasks_found))
task_found = tasks_found[0]
self.assertEqual(self._e654a0648db0_task_uuid, task_found.uuid)
self.assertEqual(self._e654a0648db0_deployment_uuid,
task_found.deployment_uuid)
self.assertEqual(consts.TaskStatus.FINISHED, task_found.status)
# NOTE(ikhudoshyn): if for all workloads success == True
self.assertFalse(task_found.pass_sla)
# NOTE(ikhudoshyn): sum of all full_durations of all workloads
self.assertEqual(142, task_found.task_duration)
# NOTE(ikhudoshyn): we have no info on validation duration in old
# schema
self.assertEqual(0, task_found.validation_duration)
self.assertEqual({}, json.loads(task_found.validation_result))
# Check subtask
subtasks_found = conn.execute(
subtask_table.select().
where(subtask_table.c.task_uuid ==
self._e654a0648db0_task_uuid)
).fetchall()
self.assertEqual(1, len(subtasks_found))
subtask_found = subtasks_found[0]
self.assertEqual(self._e654a0648db0_task_uuid,
subtask_found.task_uuid)
# NOTE(ikhudoshyn): if for all workloads success == True
self.assertFalse(subtask_found.pass_sla)
# NOTE(ikhudoshyn): sum of all full_durations of all workloads
self.assertEqual(142, subtask_found.duration)
self._e654a0648db0_subtask_uuid = subtask_found.uuid
# Check tag
tags_found = conn.execute(
tag_table.select().
where(tag_table.c.uuid == self._e654a0648db0_task_uuid)
).fetchall()
self.assertEqual(1, len(tags_found))
self.assertEqual("test_tag", tags_found[0].tag)
self.assertEqual(consts.TagType.TASK, tags_found[0].type)
# Check workload
workloads_found = conn.execute(
workload_table.select().
where(workload_table.c.task_uuid ==
self._e654a0648db0_task_uuid)
).fetchall()
self.assertEqual(1, len(workloads_found))
workload_found = workloads_found[0]
self.assertEqual(self._e654a0648db0_task_uuid,
workload_found.task_uuid)
self.assertEqual(self._e654a0648db0_subtask_uuid,
workload_found.subtask_uuid)
self.assertEqual("test_scenario", workload_found.name)
self.assertEqual(0, workload_found.position)
self.assertEqual("theRunner", workload_found.runner_type)
self.assertEqual(json.dumps({"type": "theRunner"}),
workload_found.runner)
self.assertEqual(json.dumps({"s": "S"}), workload_found.sla)
self.assertEqual(json.dumps({"a": "A"}), workload_found.args)
self.assertEqual(json.dumps({"c": "C"}), workload_found.context)
self.assertEqual(json.dumps({
"sla": [{"success": True},
{"success": False}]
}), workload_found.sla_results)
self.assertEqual(json.dumps({}), workload_found.context_execution)
self.assertEqual(42, workload_found.load_duration)
self.assertEqual(142, workload_found.full_duration)
self.assertEqual(1, workload_found.min_duration)
self.assertEqual(8, workload_found.max_duration)
self.assertEqual(3, workload_found.total_iteration_count)
self.assertEqual(1, workload_found.failed_iteration_count)
self.assertFalse(workload_found.pass_sla)
self._e654a0648db0_workload_uuid = workload_found.uuid
# Check workloadData
workloaddata_found = conn.execute(
workloaddata_table.select().
where(workloaddata_table.c.task_uuid ==
self._e654a0648db0_task_uuid)
).fetchall()
self.assertEqual(1, len(workloaddata_found))
wloaddata_found = workloaddata_found[0]
self.assertEqual(self._e654a0648db0_task_uuid,
wloaddata_found.task_uuid)
self.assertEqual(self._e654a0648db0_workload_uuid,
wloaddata_found.workload_uuid)
self.assertEqual(0, wloaddata_found.chunk_order)
self.assertEqual(0, wloaddata_found.chunk_size)
self.assertEqual(0, wloaddata_found.compressed_chunk_size)
self.assertEqual(3, wloaddata_found.iteration_count)
self.assertEqual(1, wloaddata_found.failed_iteration_count)
self.assertEqual(
json.dumps(
{
"raw": [
{"error": "e", "duration": 3},
{"duration": 1},
{"duration": 8},
]
}
), wloaddata_found.chunk_data
)
# Delete all stuff created at _pre_upgrade step
conn.execute(
tag_table.delete().
where(tag_table.c.uuid == self._e654a0648db0_task_uuid)
)
conn.execute(
workloaddata_table.delete().
where(workloaddata_table.c.task_uuid ==
self._e654a0648db0_task_uuid)
)
conn.execute(
workload_table.delete().
where(workload_table.c.task_uuid ==
self._e654a0648db0_task_uuid)
)
conn.execute(
subtask_table.delete().
where(subtask_table.c.task_uuid ==
self._e654a0648db0_task_uuid)
)
conn.execute(
task_table.delete().
where(task_table.c.uuid == self._e654a0648db0_task_uuid)
)
conn.execute(
deployment_table.delete().
where(deployment_table.c.uuid ==
self._e654a0648db0_deployment_uuid)
)
def _pre_upgrade_6ad4f426f005(self, engine):
deployment_table = db_utils.get_table(engine, "deployments")
task_table = db_utils.get_table(engine, "tasks")
task_result_table = db_utils.get_table(engine, "task_results")
with engine.connect() as conn:
# create deployment
conf = {
"admin": {"username": "admin",
"password": "passwd",
"project_name": "admin"},
"auth_url": "http://example.com:5000/v3",
"region_name": "RegionOne",
"type": "ExistingCloud"
}
deployment_status = consts.DeployStatus.DEPLOY_FINISHED
conn.execute(
deployment_table.insert(),
[{
"uuid": "my_deployment",
"name": "my_deployment",
"config": json.dumps(conf),
"enum_deployments_status": deployment_status,
"credentials": six.b(json.dumps([])),
"users": six.b(json.dumps([]))
}])
# create task
conn.execute(
task_table.insert(),
[{
"uuid": "my_task",
"deployment_uuid": "my_deployment",
"status": consts.TaskStatus.INIT,
}])
# create task result with empty data
conn.execute(
task_result_table.insert(),
[{
"task_uuid": "my_task",
"key": json.dumps({}),
"data": json.dumps({}),
}]
)
def _check_6ad4f426f005(self, engine, data):
self.assertEqual("6ad4f426f005",
api.get_backend().schema_revision(engine=engine))
deployment_table = db_utils.get_table(engine, "deployments")
task_table = db_utils.get_table(engine, "tasks")
task_result_table = db_utils.get_table(engine, "task_results")
with engine.connect() as conn:
task_results = conn.execute(task_result_table.select()).fetchall()
self.assertEqual(1, len(task_results))
task_result = task_results[0]
# check that "hooks" field added
self.assertEqual({"hooks": []}, json.loads(task_result.data))
# Remove task result
conn.execute(
task_result_table.delete().where(
task_result_table.c.id == task_result.id)
)
# Remove task
conn.execute(
task_table.delete().where(task_table.c.uuid == "my_task"))
# Remove deployment
conn.execute(
deployment_table.delete().where(
deployment_table.c.uuid == "my_deployment")
)
def _pre_upgrade_32fada9b2fde(self, engine):
self._32fada9b2fde_deployments = {
# right config which should not be changed after migration
"should-not-be-changed-1": {
"admin": {"username": "admin",
"password": "passwd",
"project_name": "admin"},
"auth_url": "http://example.com:5000/v3",
"region_name": "RegionOne",
"type": "ExistingCloud"},
# right config which should not be changed after migration
"should-not-be-changed-2": {
"admin": {"username": "admin",
"password": "passwd",
"tenant_name": "admin"},
"users": [{"username": "admin",
"password": "passwd",
"tenant_name": "admin"}],
"auth_url": "http://example.com:5000/v2.0",
"region_name": "RegionOne",
"type": "ExistingCloud"},
# not ExistingCloud config which should not be changed
"should-not-be-changed-3": {
"url": "example.com",
"type": "Something"},
# with `admin_domain_name` field
"with_admin_domain_name": {
"admin": {"username": "admin",
"password": "passwd",
"project_name": "admin",
"admin_domain_name": "admin"},
"auth_url": "http://example.com:5000/v3",
"region_name": "RegionOne",
"type": "ExistingCloud"},
}
deployment_table = db_utils.get_table(engine, "deployments")
deployment_status = consts.DeployStatus.DEPLOY_FINISHED
with engine.connect() as conn:
for deployment in self._32fada9b2fde_deployments:
conf = json.dumps(
self._32fada9b2fde_deployments[deployment])
conn.execute(
deployment_table.insert(),
[{"uuid": deployment, "name": deployment,
"config": conf,
"enum_deployments_status": deployment_status,
"credentials": six.b(json.dumps([])),
"users": six.b(json.dumps([]))
}])
def _check_32fada9b2fde(self, engine, data):
self.assertEqual("32fada9b2fde",
api.get_backend().schema_revision(engine=engine))
original_deployments = self._32fada9b2fde_deployments
deployment_table = db_utils.get_table(engine, "deployments")
with engine.connect() as conn:
deployments_found = conn.execute(
deployment_table.select()).fetchall()
for deployment in deployments_found:
# check deployment
self.assertIn(deployment.uuid, original_deployments)
self.assertIn(deployment.name, original_deployments)
config = json.loads(deployment.config)
if config != original_deployments[deployment.uuid]:
if deployment.uuid.startswith("should-not-be-changed"):
self.fail("Config of deployment '%s' is changes, but "
"should not." % deployment.uuid)
if "admin_domain_name" in deployment.config:
self.fail("Config of deployment '%s' should not "
"contain `admin_domain_name` field." %
deployment.uuid)
endpoint_type = (original_deployments[
deployment.uuid].get("endpoint_type"))
if endpoint_type in (None, "public"):
self.assertNotIn("endpoint_type", config)
else:
self.assertIn("endpoint_type", config)
self.assertEqual(endpoint_type,
config["endpoint_type"])
existing.ExistingCloud({"config": config}).validate()
else:
if not deployment.uuid.startswith("should-not-be-changed"):
self.fail("Config of deployment '%s' is not changes, "
"but should." % deployment.uuid)
# this deployment created at _pre_upgrade step is not needed
# anymore and we can remove it
conn.execute(
deployment_table.delete().where(
deployment_table.c.uuid == deployment.uuid)
)
def _pre_upgrade_484cd9413e66(self, engine):
self._484cd9413e66_deployment_uuid = "484cd9413e66-deploy"
self._484cd9413e66_verifications = [
{"total": {"time": 1.0,
"failures": 2,
"skipped": 3,
"success": 4,
"errors": 0,
"tests": 2
},
"test_cases": {"test1": {"status": "OK"},
"test2": {"status": "FAIL",
"failure": {"log": "trace"}}},
"set_name": "full"},
{"total": {"time": 2.0,
"failures": 3,
"skipped": 4,
"success": 5,
"unexpected_success": 6,
"expected_failures": 7,
"tests": 2
},
"test_cases": {"test1": {"status": "success"},
"test2": {"status": "failed", ""
"traceback": "trace"}},
"set_name": "smoke"}
]
deployment_table = db_utils.get_table(engine, "deployments")
verifications_table = db_utils.get_table(engine, "verifications")
vresults_table = db_utils.get_table(engine,
"verification_results")
deployment_status = consts.DeployStatus.DEPLOY_FINISHED
vstatus = consts.TaskStatus.FINISHED
with engine.connect() as conn:
conn.execute(
deployment_table.insert(),
[{"uuid": self._484cd9413e66_deployment_uuid,
"name": self._484cd9413e66_deployment_uuid,
"config": six.b(json.dumps([])),
"enum_deployments_status": deployment_status,
"credentials": six.b(json.dumps([])),
"users": six.b(json.dumps([]))
}])
for i in range(len(self._484cd9413e66_verifications)):
verification = self._484cd9413e66_verifications[i]
vuuid = "uuid-%s" % i
conn.execute(
verifications_table.insert(),
[{"uuid": vuuid,
"deployment_uuid":
self._484cd9413e66_deployment_uuid,
"status": vstatus,
"set_name": verification["set_name"],
"tests": verification["total"]["tests"],
"failures": verification["total"]["failures"],
"time": verification["total"]["time"],
"errors": 0,
}])
data = copy.deepcopy(verification)
data["total"]["test_cases"] = data["test_cases"]
data = data["total"]
conn.execute(
vresults_table.insert(),
[{"uuid": vuuid,
"verification_uuid": vuuid,
"data": json.dumps(data)
}])
def _check_484cd9413e66(self, engine, data):
self.assertEqual("484cd9413e66",
api.get_backend().schema_revision(engine=engine))
verifications_table = db_utils.get_table(engine, "verifications")
with engine.connect() as conn:
verifications = conn.execute(
verifications_table.select()).fetchall()
for i in range(len(verifications)):
verification_orig = self._484cd9413e66_verifications[i]
verification = verifications[i]
total = {"time": verification.tests_duration,
"failures": verification.failures,
"skipped": verification.skipped,
"success": verification.success,
"tests": verification.tests_count}
results = verification_orig["test_cases"]
old_format = "errors" in verification_orig["total"]
if old_format:
total["errors"] = 0
for test_name in results:
status = results[test_name]["status"]
if status == "OK":
status = "success"
elif status == "FAIL":
status = "fail"
results[test_name]["traceback"] = results[
test_name]["failure"].pop("log")
results[test_name].pop("failure")
results[test_name]["status"] = status
else:
uxsucess = verification.unexpected_success
total["unexpected_success"] = uxsucess
total["expected_failures"] = verification.expected_failures
self.assertEqual(verification_orig["total"], total)
self.assertEqual(results, json.loads(verification.tests))
self.assertEqual(
{"pattern": "set=%s" % verification_orig["set_name"]},
json.loads(verification.run_args))
self.assertEqual(
verification_orig["total"].get("unexpected_success", 0),
verification.unexpected_success)
self.assertEqual(
verification_orig["total"].get("expected_failures", 0),
verification.expected_failures)
conn.execute(
verifications_table.delete().where(
verifications_table.c.uuid == verification.uuid)
)
deployment_table = db_utils.get_table(engine, "deployments")
conn.execute(
deployment_table.delete().where(
deployment_table.c.uuid ==
self._484cd9413e66_deployment_uuid)
)
def _pre_upgrade_37fdbb373e8d(self, engine):
self._37fdbb373e8d_deployment_uuid = "37fdbb373e8d-deployment"
self._37fdbb373e8d_verifier_uuid = "37fdbb373e8d-verifier"
self._37fdbb373e8d_verifications_tests = [
{
"test_1[smoke, negative]": {
"name": "test_1",
"time": 2.32,
"status": "success",
"tags": ["smoke", "negative"]
},
"test_2[smoke, negative]": {
"name": "test_2",
"time": 4.32,
"status": "success",
"tags": ["smoke", "negative"]
}
},
{
"test_3[smoke, negative]": {
"name": "test_3",
"time": 6.32,
"status": "success",
"tags": ["smoke", "negative"]
},
"test_4[smoke, negative]": {
"name": "test_4",
"time": 8.32,
"status": "success",
"tags": ["smoke", "negative"]
}
}
]
deployment_table = db_utils.get_table(engine, "deployments")
verifiers_table = db_utils.get_table(engine, "verifiers")
verifications_table = db_utils.get_table(engine, "verifications")
deployment_status = consts.DeployStatus.DEPLOY_FINISHED
with engine.connect() as conn:
conn.execute(
deployment_table.insert(),
[{"uuid": self._37fdbb373e8d_deployment_uuid,
"name": self._37fdbb373e8d_deployment_uuid,
"config": six.b(json.dumps([])),
"enum_deployments_status": deployment_status,
"credentials": six.b(json.dumps([])),
"users": six.b(json.dumps([]))
}])
conn.execute(
verifiers_table.insert(),
[{"uuid": self._37fdbb373e8d_verifier_uuid,
"name": self._37fdbb373e8d_verifier_uuid,
"type": "some-type",
"status": consts.VerifierStatus.INSTALLED
}])
for i in range(len(self._37fdbb373e8d_verifications_tests)):
tests = self._37fdbb373e8d_verifications_tests[i]
conn.execute(
verifications_table.insert(),
[{"uuid": "verification-uuid-%s" % i,
"deployment_uuid":
self._37fdbb373e8d_deployment_uuid,
"verifier_uuid": self._37fdbb373e8d_verifier_uuid,
"status": consts.VerificationStatus.FINISHED,
"tests": json.dumps(tests)
}])
def _check_37fdbb373e8d(self, engine, data):
self.assertEqual("37fdbb373e8d",
api.get_backend().schema_revision(engine=engine))
verifications_table = db_utils.get_table(engine, "verifications")
with engine.connect() as conn:
verifications = conn.execute(
verifications_table.select()).fetchall()
self.assertEqual(len(verifications),
len(self._37fdbb373e8d_verifications_tests))
for i in range(len(verifications)):
v = verifications[i]
updated_tests = json.loads(v.tests)
expected_tests = self._37fdbb373e8d_verifications_tests[i]
for test in expected_tests.values():
duration = test.pop("time")
test["duration"] = duration
self.assertEqual(expected_tests, updated_tests)
conn.execute(
verifications_table.delete().where(
verifications_table.c.uuid == v.uuid)
)
deployment_table = db_utils.get_table(engine, "deployments")
conn.execute(
deployment_table.delete().where(
deployment_table.c.uuid ==
self._37fdbb373e8d_deployment_uuid)
)
def _pre_upgrade_a6f364988fc2(self, engine):
self._a6f364988fc2_tags = [
{
"uuid": "uuid-1",
"type": "task",
"tag": "tag-1"
},
{
"uuid": "uuid-2",
"type": "subtask",
"tag": "tag-2"
},
{
"uuid": "uuid-3",
"type": "task",
"tag": "tag-3"
}
]
tags_table = db_utils.get_table(engine, "tags")
with engine.connect() as conn:
for t in self._a6f364988fc2_tags:
conn.execute(
tags_table.insert(),
[{
"uuid": t["uuid"],
"enum_tag_types": t["type"],
"type": t["type"],
"tag": t["tag"]
}])
def _check_a6f364988fc2(self, engine, data):
self.assertEqual("a6f364988fc2",
api.get_backend().schema_revision(engine=engine))
tags_table = db_utils.get_table(engine, "tags")
with engine.connect() as conn:
tags = conn.execute(tags_table.select()).fetchall()
self.assertEqual(len(tags), len(self._a6f364988fc2_tags))
for i in range(len(tags)):
for k in ("uuid", "type", "tag"):
self.assertEqual(self._a6f364988fc2_tags[i][k], tags[i][k])
conn.execute(
tags_table.delete().where(
tags_table.c.uuid == tags[i].uuid))
def _pre_upgrade_f33f4610dcda(self, engine):
self._f33f4610dcda_deployment_uuid = "f33f4610dcda-deployment"
self._f33f4610dcda_verifier_uuid = "f33f4610dcda-verifier"
self._f33f4610dcda_verifications = [
{"status": "init", "failures": 0, "unexpected_success": 0},
{"status": "running", "failures": 0, "unexpected_success": 0},
{"status": "finished", "failures": 0, "unexpected_success": 0},
{"status": "finished", "failures": 1, "unexpected_success": 0,
"new_status": "failed"},
{"status": "finished", "failures": 1, "unexpected_success": 1,
"new_status": "failed"},
{"status": "finished", "failures": 0, "unexpected_success": 1,
"new_status": "failed"},
{"status": "failed", "failures": 0, "unexpected_success": 0,
"new_status": "crashed"},
]
deployment_table = db_utils.get_table(engine, "deployments")
verifiers_table = db_utils.get_table(engine, "verifiers")
verifications_table = db_utils.get_table(engine, "verifications")
deployment_status = consts.DeployStatus.DEPLOY_FINISHED
with engine.connect() as conn:
conn.execute(
deployment_table.insert(),
[{"uuid": self._f33f4610dcda_deployment_uuid,
"name": self._f33f4610dcda_deployment_uuid,
"config": six.b(json.dumps([])),
"enum_deployments_status": deployment_status,
"credentials": six.b(json.dumps([])),
"users": six.b(json.dumps([]))
}])
conn.execute(
verifiers_table.insert(),
[{"uuid": self._f33f4610dcda_verifier_uuid,
"name": self._f33f4610dcda_verifier_uuid,
"type": "some-type",
"status": consts.VerifierStatus.INSTALLED
}])
for i in range(len(self._f33f4610dcda_verifications)):
v = self._f33f4610dcda_verifications[i]
conn.execute(
verifications_table.insert(),
[{"uuid": "verification-uuid-%s" % i,
"deployment_uuid": self._f33f4610dcda_deployment_uuid,
"verifier_uuid": self._f33f4610dcda_verifier_uuid,
"status": v["status"],
"failures": v["failures"],
"unexpected_success": v["unexpected_success"]
}])
def _check_f33f4610dcda(self, engine, data):
self.assertEqual("f33f4610dcda",
api.get_backend().schema_revision(engine=engine))
verifications_table = db_utils.get_table(engine, "verifications")
with engine.connect() as conn:
verifications = conn.execute(
verifications_table.select()).fetchall()
self.assertEqual(len(verifications),
len(self._f33f4610dcda_verifications))
for i in range(len(verifications)):
if "new_status" in self._f33f4610dcda_verifications[i]:
self.assertEqual(
self._f33f4610dcda_verifications[i]["new_status"],
verifications[i].status)
conn.execute(
verifications_table.delete().where(
verifications_table.c.uuid == verifications[i].uuid)
)
deployment_table = db_utils.get_table(engine, "deployments")
conn.execute(
deployment_table.delete().where(
deployment_table.c.uuid ==
self._f33f4610dcda_deployment_uuid)
)
def _pre_upgrade_4ef544102ba7(self, engine):
self._4ef544102ba7_deployment_uuid = "4ef544102ba7-deploy"
self.tasks = {
"should-not-be-changed-1": {
"uuid": "should-not-be-changed-1",
"deployment_uuid": self._4ef544102ba7_deployment_uuid,
"validation_result": {
"etype": "SomeCls",
"msg": "msg",
"trace": "Traceback (most recent call last):\n"
"File some1.py, line ...\n"
"File some2.py, line ...\nSomeCls: msg"},
"status": "finished"},
"should-be-changed-1": {
"uuid": "should-be-changed-1",
"deployment_uuid": self._4ef544102ba7_deployment_uuid,
"validation_result": {},
"status": "failed"},
"should-be-changed-2": {
"uuid": "should-be-changed-2",
"deployment_uuid": self._4ef544102ba7_deployment_uuid,
"validation_result": {},
"status": "verifying"},
}
deployment_table = db_utils.get_table(engine, "deployments")
with engine.connect() as conn:
conn.execute(
deployment_table.insert(),
[{"uuid": self._4ef544102ba7_deployment_uuid,
"name": self._4ef544102ba7_deployment_uuid,
"config": six.b(json.dumps([])),
"enum_deployments_status":
consts.DeployStatus.DEPLOY_FINISHED,
"credentials": six.b(json.dumps([])),
"users": six.b(json.dumps([]))
}])
task_table = db_utils.get_table(engine, "tasks")
with engine.connect() as conn:
for task in self.tasks:
conn.execute(
task_table.insert(),
[{
"deployment_uuid": self.tasks[task][
"deployment_uuid"],
"status": self.tasks[task]["status"],
"validation_result": json.dumps(
self.tasks[task]["validation_result"]),
"uuid": self.tasks[task]["uuid"]
}])
subtask_table = db_utils.get_table(engine, "subtasks")
with engine.connect() as conn:
for task in self.tasks:
conn.execute(
subtask_table.insert(),
[{
"task_uuid": self.tasks[task]["uuid"],
"status": consts.SubtaskStatus.RUNNING,
"context": json.dumps({}),
"sla": json.dumps({}),
"run_in_parallel": False,
"uuid": "subtask_" + self.tasks[task]["uuid"]
}])
def _check_4ef544102ba7(self, engine, data):
self.assertEqual("4ef544102ba7",
api.get_backend().schema_revision(engine=engine))
org_tasks = self.tasks
task_table = db_utils.get_table(engine, "tasks")
subtask_table = db_utils.get_table(engine, "subtasks")
with engine.connect() as conn:
subtasks_found = conn.execute(
subtask_table.select()).fetchall()
for subtask in subtasks_found:
conn.execute(
subtask_table.delete().where(
subtask_table.c.id == subtask.id)
)
with engine.connect() as conn:
tasks_found = conn.execute(
task_table.select()).fetchall()
self.assertEqual(3, len(tasks_found))
for task in tasks_found:
self.assertIn("uuid", task)
self.assertIn("status", task)
if task.status != org_tasks[task.uuid]["status"]:
if task.uuid.startswith("should-not-be-changed"):
self.fail("Config of deployment '%s' is changes, but "
"should not." % task.uuid)
if task.status != "crashed" and task.uuid == (
"should-be-changed-1"):
self.fail("Task '%s' status should be changed to "
"crashed." % task.uuid)
if task.status != "validating" and task.uuid == (
"should-be-changed-2"):
self.fail("Task '%s' status should be changed to "
"validating." % task.uuid)
else:
if not task.uuid.startswith("should-not-be-changed"):
self.fail("Config of deployment '%s' is not changes, "
"but should." % task.uuid)
conn.execute(
task_table.delete().where(
task_table.c.id == task.id)
)
deployment_table = db_utils.get_table(engine, "deployments")
conn.execute(
deployment_table.delete().where(
deployment_table.c.uuid ==
self._4ef544102ba7_deployment_uuid)
)
def _pre_upgrade_92aaaa2a6bb3(self, engine):
self._92aaaa2a6bb3_deployments = [
("1-cred", [["openstack", {"foo": "bar"}]]),
("2-cred", [["openstack", {"foo": "bar1"}],
["openstack", {"foo": "bar2"}]]),
("multi-cred", [["spam", {"foo": "bar1"}],
["eggs", {"foo": "bar2"}]]),
]
deployment_table = db_utils.get_table(engine, "deployments")
deployment_status = consts.DeployStatus.DEPLOY_FINISHED
with engine.connect() as conn:
for deployment, creds in self._92aaaa2a6bb3_deployments:
conn.execute(
deployment_table.insert(),
[{"uuid": deployment, "name": deployment,
"config": json.dumps({}),
"enum_deployments_status": deployment_status,
"credentials": pickle.dumps(creds),
}])
def _check_92aaaa2a6bb3(self, engine, data):
expected_credentials = [
("1-cred", {"openstack": [{"foo": "bar"}]}),
("2-cred", {"openstack": [{"foo": "bar1"},
{"foo": "bar2"}]}),
("multi-cred", {"spam": [{"foo": "bar1"}],
"eggs": [{"foo": "bar2"}]}),
]
deployment_table = db_utils.get_table(engine, "deployments")
with engine.connect() as conn:
for deployment, expected_creds in expected_credentials:
dep_obj = conn.execute(
deployment_table.select().where(
deployment_table.c.uuid == deployment)).fetchone()
self.assertEqual(
expected_creds, json.loads(dep_obj.credentials))
conn.execute(
deployment_table.delete().where(
deployment_table.c.uuid == deployment))
def _pre_upgrade_35fe16d4ab1c(self, engine):
deployment_table = db_utils.get_table(engine, "deployments")
task_table = db_utils.get_table(engine, "tasks")
subtask_table = db_utils.get_table(engine, "subtasks")
workload_table = db_utils.get_table(engine, "workloads")
deployment_uuid = str(uuid.uuid4())
self._35fe16d4ab1c_task_uuid = str(uuid.uuid4())
self._35fe16d4ab1c_subtasks = {
str(uuid.uuid4()): [
{"uuid": str(uuid.uuid4()),
"pass_sla": False,
"load_duration": 1},
{"uuid": str(uuid.uuid4()),
"pass_sla": False,
"load_duration": 2.6}
],
str(uuid.uuid4()): [
{"uuid": str(uuid.uuid4()),
"pass_sla": True,
"load_duration": 3},
{"uuid": str(uuid.uuid4()),
"pass_sla": False,
"load_duration": 7}
]
}
with engine.connect() as conn:
conn.execute(
deployment_table.insert(),
[{
"uuid": deployment_uuid,
"name": str(uuid.uuid4()),
"config": "{}",
"enum_deployments_status": consts.DeployStatus.DEPLOY_INIT,
"credentials": six.b(json.dumps([])),
"users": six.b(json.dumps([]))
}]
)
conn.execute(
task_table.insert(),
[{
"uuid": self._35fe16d4ab1c_task_uuid,
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"status": consts.TaskStatus.FINISHED,
"validation_result": six.b(json.dumps({})),
"deployment_uuid": deployment_uuid
}]
)
for subtask_id, workloads in self._35fe16d4ab1c_subtasks.items():
conn.execute(
subtask_table.insert(),
[{
"uuid": subtask_id,
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"task_uuid": self._35fe16d4ab1c_task_uuid,
"context": six.b(json.dumps([])),
"sla": six.b(json.dumps([])),
"run_in_parallel": False
}]
)
for workload in workloads:
conn.execute(
workload_table.insert(),
[{
"uuid": workload["uuid"],
"name": "foo",
"task_uuid": self._35fe16d4ab1c_task_uuid,
"subtask_uuid": subtask_id,
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"position": 0,
"runner": "",
"runner_type": "",
"context": "",
"context_execution": "",
"statistics": "",
"hooks": "",
"sla": "",
"sla_results": "",
"args": "",
"load_duration": workload["load_duration"],
"pass_sla": workload["pass_sla"]
}]
)
def _check_35fe16d4ab1c(self, engine, data):
deployment_table = db_utils.get_table(engine, "deployments")
task_table = db_utils.get_table(engine, "tasks")
subtask_table = db_utils.get_table(engine, "subtasks")
workload_table = db_utils.get_table(engine, "workloads")
with engine.connect() as conn:
task_id = self._35fe16d4ab1c_task_uuid
task_obj = conn.execute(
task_table.select().where(
task_table.c.uuid == task_id)).fetchone()
self.assertFalse(task_obj.pass_sla)
subtask_duration = dict(
[(k, sum([w["load_duration"] for w in v]))
for k, v in self._35fe16d4ab1c_subtasks.items()])
self.assertEqual(sum(subtask_duration.values()),
task_obj.task_duration)
for subtask_id, workloads in self._35fe16d4ab1c_subtasks.items():
subtask_obj = conn.execute(
subtask_table.select().where(
subtask_table.c.uuid == subtask_id)).fetchone()
self.assertFalse(subtask_obj.pass_sla)
self.assertEqual(sum([w["load_duration"] for w in workloads]),
subtask_obj.duration)
conn.execute(
workload_table.delete().where(
workload_table.c.subtask_uuid == subtask_id))
conn.execute(
subtask_table.delete().where(
subtask_table.c.uuid == subtask_id))
conn.execute(
task_table.delete().where(
task_table.c.uuid == task_obj.uuid))
conn.execute(
deployment_table.delete().where(
deployment_table.c.uuid == task_obj.deployment_uuid))
def _pre_upgrade_fab4f4f31f8a(self, engine):
deployment_table = db_utils.get_table(engine, "deployments")
task_table = db_utils.get_table(engine, "tasks")
subtask_table = db_utils.get_table(engine, "subtasks")
workload_table = db_utils.get_table(engine, "workloads")
wdata_table = db_utils.get_table(engine, "workloaddata")
self._fab4f4f31f8a_deployment_uuid = str(uuid.uuid4())
task_uuid = str(uuid.uuid4())
self._fab4f4f31f8a_subtask = str(uuid.uuid4())
self._fab4f4f31f8a_workloads = [
{"uuid": str(uuid.uuid4()),
"start_time": 0.0,
"data": [{"timestamp": 0,
# deprecated output
"scenario_output": {"data": {1: 2}},
"duration": 3,
"idle_duration": 0,
"error": None,
# old format of atomics
"atomic_actions": {"foo": 3}}],
"statistics": {"durations": {
"atomics": [{"children": [],
"count_per_iteration": 1,
"data": {"90%ile": 3.0,
"95%ile": 3.0,
"avg": 3.0,
"iteration_count": 1,
"max": 3.0,
"median": 3.0,
"min": 3.0,
"success": "100.0%"},
"display_name": "foo",
"name": "foo"}],
"total": {
"display_name": "total",
"name": "total",
"count_per_iteration": 1,
"data": {"90%ile": 3.0,
"95%ile": 3.0,
"avg": 3.0,
"iteration_count": 1,
"max": 3.0,
"median": 3.0,
"min": 3.0,
"success": "100.0%"},
"children": [
{"display_name": "duration",
"name": "duration",
"children": [],
"count_per_iteration": 1,
"data": {"90%ile": 3.0,
"95%ile": 3.0,
"avg": 3.0,
"iteration_count": 1,
"max": 3.0,
"median": 3.0,
"min": 3.0,
"success": "100.0%"}},
{"display_name": "idle_duration",
"name": "idle_duration",
"children": [],
"count_per_iteration": 1,
"data": {"90%ile": 0.0,
"95%ile": 0.0,
"avg": 0.0,
"iteration_count": 1,
"max": 0.0,
"median": 0.0,
"min": 0.0,
"success": "100.0%"}}]}}}},
{"uuid": str(uuid.uuid4()),
"start_time": 1.0,
"data": [{"timestamp": 1, "output": {},
"duration": 5,
"idle_duration": 0,
"error": None,
"atomic_actions": [
{"name": "foo", "started_at": 2,
"finished_at": 3, "children": []},
{"name": "foo", "started_at": 3,
"finished_at": 5, "children": []}]},
{"timestamp": 6,
"output": {},
"duration": 4,
"idle_duration": 0,
"error": None,
"atomic_actions": [
{"name": "foo", "started_at": 6,
"finished_at": 9, "children": []},
{"name": "foo", "started_at": 9,
"finished_at": 10, "children": []}]}],
"statistics": {"durations": {
"atomics": [{"display_name": "foo (x2)",
"name": "foo",
"children": [],
"count_per_iteration": 2,
"data": {"90%ile": 3.9,
"95%ile": 3.95,
"avg": 3.5,
"iteration_count": 2,
"max": 4.0,
"median": 3.5,
"min": 3.0,
"success": "100.0%"}}],
"total": {
"display_name": "total",
"name": "total",
"count_per_iteration": 1,
"data": {"90%ile": 4.9,
"95%ile": 4.95,
"avg": 4.5,
"iteration_count": 2,
"max": 5.0,
"median": 4.5,
"min": 4.0,
"success": "100.0%"},
"children": [
{
"display_name": "duration",
"name": "duration",
"children": [],
"count_per_iteration": 1,
"data": {"90%ile": 4.9,
"95%ile": 4.95,
"avg": 4.5,
"iteration_count": 2,
"max": 5.0,
"median": 4.5,
"min": 4.0,
"success": "100.0%"}},
{
"display_name": "idle_duration",
"name": "idle_duration",
"children": [],
"count_per_iteration": 1,
"data": {"90%ile": 0.0,
"95%ile": 0.0,
"avg": 0.0,
"iteration_count": 2,
"max": 0.0,
"median": 0.0,
"min": 0.0,
"success": "100.0%"}
}
]
}
}}}
]
with engine.connect() as conn:
conn.execute(
deployment_table.insert(),
[{
"uuid": self._fab4f4f31f8a_deployment_uuid,
"name": str(uuid.uuid4()),
"config": "{}",
"enum_deployments_status": consts.DeployStatus.DEPLOY_INIT,
"credentials": six.b(json.dumps([])),
"users": six.b(json.dumps([]))
}]
)
conn.execute(
task_table.insert(),
[{
"uuid": task_uuid,
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"status": consts.TaskStatus.FINISHED,
"validation_result": six.b(json.dumps({})),
"deployment_uuid": self._fab4f4f31f8a_deployment_uuid
}]
)
conn.execute(
subtask_table.insert(),
[{
"uuid": self._fab4f4f31f8a_subtask,
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"task_uuid": task_uuid,
"context": six.b(json.dumps([])),
"sla": six.b(json.dumps([])),
"run_in_parallel": False
}]
)
for workload in self._fab4f4f31f8a_workloads:
conn.execute(
workload_table.insert(),
[{
"uuid": workload["uuid"],
"name": "foo",
"task_uuid": task_uuid,
"subtask_uuid": self._fab4f4f31f8a_subtask,
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"position": 0,
"runner": "",
"runner_type": "",
"context": "",
"context_execution": "",
"statistics": "",
"hooks": "",
"sla": "",
"sla_results": "",
"args": "",
"load_duration": 0,
"pass_sla": True
}]
)
conn.execute(
wdata_table.insert(),
[{
"uuid": str(uuid.uuid4()),
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"started_at": timeutils.utcnow(),
"finished_at": timeutils.utcnow(),
"task_uuid": task_uuid,
"workload_uuid": workload["uuid"],
"chunk_order": 0,
"iteration_count": 0,
"failed_iteration_count": 0,
"chunk_size": 0,
"compressed_chunk_size": 0,
"chunk_data": json.dumps({"raw": workload["data"]})
}]
)
def _check_fab4f4f31f8a(self, engine, data):
deployment_table = db_utils.get_table(engine, "deployments")
task_table = db_utils.get_table(engine, "tasks")
subtask_table = db_utils.get_table(engine, "subtasks")
workload_table = db_utils.get_table(engine, "workloads")
wdata_table = db_utils.get_table(engine, "workloaddata")
task_uuid = None
with engine.connect() as conn:
subtask_id = self._fab4f4f31f8a_subtask
for workload in conn.execute(workload_table.select().where(
workload_table.c.subtask_uuid == subtask_id)).fetchall():
if task_uuid is None:
task_uuid = workload.task_uuid
original = [w for w in self._fab4f4f31f8a_workloads
if w["uuid"] == workload.uuid][0]
if workload.start_time is None:
start_time = None
else:
start_time = workload.start_time / 1000000.0
self.assertEqual(original["start_time"], start_time)
self.assertEqual(original["statistics"],
json.loads(workload.statistics))
wuuid = workload.uuid
for wdata in conn.execute(wdata_table.select().where(
wdata_table.c.workload_uuid == wuuid)).fetchall():
for iter in json.loads(wdata.chunk_data)["raw"]:
self.assertNotIn("scenario_output", iter)
self.assertIn("output", iter)
self.assertIsInstance(iter["atomic_actions"], list)
conn.execute(
wdata_table.delete().where(
wdata_table.c.workload_uuid == workload.uuid))
conn.execute(
workload_table.delete().where(
workload_table.c.uuid == workload.uuid))
conn.execute(
subtask_table.delete().where(
subtask_table.c.uuid == subtask_id))
conn.execute(
task_table.delete().where(task_table.c.uuid == task_uuid))
deployment_uuid = self._fab4f4f31f8a_deployment_uuid
conn.execute(
deployment_table.delete().where(
deployment_table.c.uuid == deployment_uuid))
def _pre_upgrade_7948b83229f6(self, engine):
deployment_table = db_utils.get_table(engine, "deployments")
task_table = db_utils.get_table(engine, "tasks")
subtask_table = db_utils.get_table(engine, "subtasks")
workload_table = db_utils.get_table(engine, "workloads")
wdata_table = db_utils.get_table(engine, "workloaddata")
self._7948b83229f6_deployment_uuid = str(uuid.uuid4())
self._7948b83229f6_task_uuid = str(uuid.uuid4())
subtask_uuid = str(uuid.uuid4())
self._7948b83229f6_workloads = {
str(uuid.uuid4()): {"preprocessed": 1, "expected": 1},
str(uuid.uuid4()): {"preprocessed": 0, "expected": None},
str(uuid.uuid4()): {"preprocessed": 0, "expected": 0,
"wdata": True},
str(uuid.uuid4()): {"preprocessed": -1, "expected": None}}
with engine.connect() as conn:
conn.execute(
deployment_table.insert(),
[{
"uuid": self._7948b83229f6_deployment_uuid,
"name": str(uuid.uuid4()),
"config": "{}",
"enum_deployments_status": consts.DeployStatus.DEPLOY_INIT,
"credentials": six.b(json.dumps([])),
"users": six.b(json.dumps([]))
}]
)
conn.execute(
task_table.insert(),
[{
"uuid": self._7948b83229f6_task_uuid,
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"status": consts.TaskStatus.FINISHED,
"validation_result": six.b(json.dumps({})),
"deployment_uuid": self._7948b83229f6_deployment_uuid
}]
)
conn.execute(
subtask_table.insert(),
[{
"uuid": subtask_uuid,
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"task_uuid": self._7948b83229f6_task_uuid,
"context": six.b(json.dumps([])),
"sla": six.b(json.dumps([])),
"run_in_parallel": False
}]
)
for w_uuid, workload in self._7948b83229f6_workloads.items():
conn.execute(
workload_table.insert(),
[{
"uuid": w_uuid,
"name": "foo",
"task_uuid": self._7948b83229f6_task_uuid,
"subtask_uuid": subtask_uuid,
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"position": 0,
"runner": "",
"runner_type": "",
"context": "",
"context_execution": "",
"statistics": "",
"hooks": "",
"sla": "",
"sla_results": "",
"args": "",
"load_duration": 0,
"pass_sla": True,
"min_duration": workload["preprocessed"],
"max_duration": workload["preprocessed"]
}]
)
if workload.get("wdata", False):
conn.execute(
wdata_table.insert(),
[{
"uuid": str(uuid.uuid4()),
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"started_at": timeutils.utcnow(),
"finished_at": timeutils.utcnow(),
"task_uuid": self._7948b83229f6_task_uuid,
"workload_uuid": w_uuid,
"chunk_order": 0,
"iteration_count": 0,
"failed_iteration_count": 0,
"chunk_size": 0,
"compressed_chunk_size": 0,
"chunk_data": six.b(json.dumps([]))
}]
)
def _check_7948b83229f6(self, engine, data):
deployment_table = db_utils.get_table(engine, "deployments")
task_table = db_utils.get_table(engine, "tasks")
subtask_table = db_utils.get_table(engine, "subtasks")
workload_table = db_utils.get_table(engine, "workloads")
wdata_table = db_utils.get_table(engine, "workloaddata")
subtask_uuid = None
with engine.connect() as conn:
task_uuid = self._7948b83229f6_task_uuid
for workload in conn.execute(workload_table.select().where(
workload_table.c.task_uuid == task_uuid)).fetchall():
if subtask_uuid is None:
subtask_uuid = workload.subtask_uuid
if workload.uuid not in self._7948b83229f6_workloads:
self.fail("Unknown workload found for 7948b83229f6 "
"migration.")
original = self._7948b83229f6_workloads[workload.uuid]
self.assertEqual(original["expected"],
workload.min_duration)
self.assertEqual(original["expected"],
workload.max_duration)
if original.get("wdata", False):
conn.execute(
wdata_table.delete().where(
wdata_table.c.workload_uuid == workload.uuid))
conn.execute(
workload_table.delete().where(
workload_table.c.uuid == workload.uuid))
conn.execute(
subtask_table.delete().where(
subtask_table.c.uuid == subtask_uuid))
conn.execute(
task_table.delete().where(task_table.c.uuid == task_uuid))
deployment_uuid = self._7948b83229f6_deployment_uuid
conn.execute(
deployment_table.delete().where(
deployment_table.c.uuid == deployment_uuid))
def _pre_upgrade_046a38742e89(self, engine):
deployment_table = db_utils.get_table(engine, "deployments")
task_table = db_utils.get_table(engine, "tasks")
subtask_table = db_utils.get_table(engine, "subtasks")
workload_table = db_utils.get_table(engine, "workloads")
self._046a38742e89_deployment_uuid = str(uuid.uuid4())
self._046a38742e89_task_uuid = str(uuid.uuid4())
subtask_uuid = str(uuid.uuid4())
workloads = [
{
"runner": {"type": "constant",
"times": 1000}},
{
"runner": {"type": "rps",
"rps": 300},
"hooks": [
{
"config": {"args": {"arg1": "v1"},
"description": "descr",
"name": "foo",
"trigger": {"name": "bar",
"args": {"arg2": "v2"}}}}
]
}
]
with engine.connect() as conn:
conn.execute(
deployment_table.insert(),
[{
"uuid": self._046a38742e89_deployment_uuid,
"name": str(uuid.uuid4()),
"config": "{}",
"enum_deployments_status": consts.DeployStatus.DEPLOY_INIT,
"credentials": six.b(json.dumps([])),
"users": six.b(json.dumps([]))
}]
)
conn.execute(
task_table.insert(),
[{
"uuid": self._046a38742e89_task_uuid,
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"status": consts.TaskStatus.FINISHED,
"validation_result": six.b(json.dumps({})),
"deployment_uuid": self._046a38742e89_deployment_uuid
}]
)
conn.execute(
subtask_table.insert(),
[{
"uuid": subtask_uuid,
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"task_uuid": self._046a38742e89_task_uuid,
"context": six.b(json.dumps([])),
"sla": six.b(json.dumps([])),
"run_in_parallel": False
}]
)
for workload in workloads:
conn.execute(
workload_table.insert(),
[{
"uuid": str(uuid.uuid4()),
"name": "foo",
"task_uuid": self._046a38742e89_task_uuid,
"subtask_uuid": subtask_uuid,
"created_at": timeutils.utcnow(),
"updated_at": timeutils.utcnow(),
"position": 0,
"runner": json.dumps(workload["runner"]),
"runner_type": "",
"context": "",
"context_execution": "",
"statistics": "",
"hooks": json.dumps(workload.get("hooks", "")),
"sla": "",
"sla_results": "",
"args": "",
"load_duration": 0,
"pass_sla": True,
"min_duration": 0,
"max_duration": 1
}]
)
def _check_046a38742e89(self, engine, data):
deployment_table = db_utils.get_table(engine, "deployments")
task_table = db_utils.get_table(engine, "tasks")
subtask_table = db_utils.get_table(engine, "subtasks")
workload_table = db_utils.get_table(engine, "workloads")
subtask_uuid = None
with engine.connect() as conn:
task_uuid = self._046a38742e89_task_uuid
for workload in conn.execute(workload_table.select().where(
workload_table.c.task_uuid == task_uuid)).fetchall():
if subtask_uuid is None:
subtask_uuid = workload.subtask_uuid
runner = json.loads(workload.runner)
self.assertNotIn("type", runner)
hooks = json.loads(workload.hooks)
if hooks:
for hook in hooks:
hook_cfg = hook["config"]
self.assertEqual(2, len(hook_cfg["action"]))
self.assertEqual(2, len(hook_cfg["trigger"]))
conn.execute(
workload_table.delete().where(
workload_table.c.uuid == workload.uuid))
conn.execute(
subtask_table.delete().where(
subtask_table.c.uuid == subtask_uuid))
conn.execute(
task_table.delete().where(task_table.c.uuid == task_uuid))
deployment_uuid = self._046a38742e89_deployment_uuid
conn.execute(
deployment_table.delete().where(
deployment_table.c.uuid == deployment_uuid))