2733 lines
114 KiB
Python
2733 lines
114 KiB
Python
# vim: tabstop=4 shiftwidth=4 softtabstop=4
|
|
|
|
# Copyright 2010-2011 OpenStack Foundation
|
|
# Copyright 2012-2013 IBM Corp.
|
|
# All Rights Reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
"""
|
|
Tests for database migrations. This test case reads the configuration
|
|
file test_migrations.conf for database connection settings
|
|
to use in the tests. For each connection found in the config file,
|
|
the test case runs a series of test cases to ensure that migrations work
|
|
properly both upgrading and downgrading, and that no data loss occurs
|
|
if possible.
|
|
|
|
There are also "opportunistic" tests for both mysql and postgresql in here,
|
|
which allows testing against all 3 databases (sqlite in memory, mysql, pg) in
|
|
a properly configured unit test environment.
|
|
|
|
For the opportunistic testing you need to set up a db named 'openstack_citest'
|
|
with user 'openstack_citest' and password 'openstack_citest' on localhost.
|
|
The test will then use that db and u/p combo to run the tests.
|
|
|
|
For postgres on Ubuntu this can be done with the following commands:
|
|
|
|
sudo -u postgres psql
|
|
postgres=# create user openstack_citest with createdb login password
|
|
'openstack_citest';
|
|
postgres=# create database openstack_citest with owner openstack_citest;
|
|
|
|
"""
|
|
|
|
import collections
|
|
import commands
|
|
import ConfigParser
|
|
import datetime
|
|
import glob
|
|
import operator
|
|
import os
|
|
import urlparse
|
|
import uuid
|
|
|
|
from migrate.versioning import repository
|
|
import netaddr
|
|
import sqlalchemy
|
|
from sqlalchemy.dialects import postgresql
|
|
from sqlalchemy.dialects import sqlite
|
|
import sqlalchemy.exc
|
|
|
|
from nova.db.sqlalchemy import api as db
|
|
import nova.db.sqlalchemy.migrate_repo
|
|
from nova.db.sqlalchemy import utils as db_utils
|
|
from nova.openstack.common.gettextutils import _
|
|
from nova.openstack.common import log as logging
|
|
from nova.openstack.common import timeutils
|
|
from nova.openstack.common import uuidutils
|
|
from nova import test
|
|
from nova.tests import matchers
|
|
from nova import utils
|
|
import nova.virt.baremetal.db.sqlalchemy.migrate_repo
|
|
|
|
|
|
LOG = logging.getLogger(__name__)
|
|
|
|
|
|
def _get_connect_string(backend, user, passwd, database):
|
|
"""
|
|
Try to get a connection with a very specific set of values, if we get
|
|
these then we'll run the tests, otherwise they are skipped
|
|
"""
|
|
if backend == "postgres":
|
|
backend = "postgresql+psycopg2"
|
|
elif backend == "mysql":
|
|
backend = "mysql+mysqldb"
|
|
else:
|
|
raise Exception("Unrecognized backend: '%s'" % backend)
|
|
|
|
return ("%s://%s:%s@localhost/%s" % (backend, user, passwd, database))
|
|
|
|
|
|
def _is_backend_avail(backend, user, passwd, database):
|
|
try:
|
|
connect_uri = _get_connect_string(backend, user, passwd, database)
|
|
engine = sqlalchemy.create_engine(connect_uri)
|
|
connection = engine.connect()
|
|
except Exception:
|
|
# intentionally catch all to handle exceptions even if we don't
|
|
# have any backend code loaded.
|
|
return False
|
|
else:
|
|
connection.close()
|
|
engine.dispose()
|
|
return True
|
|
|
|
|
|
def _have_mysql(user, passwd, database):
|
|
present = os.environ.get('NOVA_TEST_MYSQL_PRESENT')
|
|
if present is None:
|
|
return _is_backend_avail('mysql', user, passwd, database)
|
|
return present.lower() in ('', 'true')
|
|
|
|
|
|
def _have_postgresql(user, passwd, database):
|
|
present = os.environ.get('NOVA_TEST_POSTGRESQL_PRESENT')
|
|
if present is None:
|
|
return _is_backend_avail('postgres', user, passwd, database)
|
|
return present.lower() in ('', 'true')
|
|
|
|
|
|
def get_mysql_connection_info(conn_pieces):
|
|
database = conn_pieces.path.strip('/')
|
|
loc_pieces = conn_pieces.netloc.split('@')
|
|
host = loc_pieces[1]
|
|
auth_pieces = loc_pieces[0].split(':')
|
|
user = auth_pieces[0]
|
|
password = ""
|
|
if len(auth_pieces) > 1:
|
|
if auth_pieces[1].strip():
|
|
password = "-p\"%s\"" % auth_pieces[1]
|
|
|
|
return (user, password, database, host)
|
|
|
|
|
|
def get_pgsql_connection_info(conn_pieces):
|
|
database = conn_pieces.path.strip('/')
|
|
loc_pieces = conn_pieces.netloc.split('@')
|
|
host = loc_pieces[1]
|
|
|
|
auth_pieces = loc_pieces[0].split(':')
|
|
user = auth_pieces[0]
|
|
password = ""
|
|
if len(auth_pieces) > 1:
|
|
password = auth_pieces[1].strip()
|
|
|
|
return (user, password, database, host)
|
|
|
|
|
|
class CommonTestsMixIn(object):
|
|
"""These tests are shared between TestNovaMigrations and
|
|
TestBaremetalMigrations.
|
|
|
|
BaseMigrationTestCase is effectively an abstract class, meant to be derived
|
|
from and not directly tested against; that's why these `test_` methods need
|
|
to be on a Mixin, so that they won't be picked up as valid tests for
|
|
BaseMigrationTestCase.
|
|
"""
|
|
def test_walk_versions(self):
|
|
for key, engine in self.engines.items():
|
|
self._walk_versions(engine, self.snake_walk)
|
|
|
|
def test_mysql_opportunistically(self):
|
|
self._test_mysql_opportunistically()
|
|
|
|
def test_mysql_connect_fail(self):
|
|
"""
|
|
Test that we can trigger a mysql connection failure and we fail
|
|
gracefully to ensure we don't break people without mysql
|
|
"""
|
|
if _is_backend_avail('mysql', "openstack_cifail", self.PASSWD,
|
|
self.DATABASE):
|
|
self.fail("Shouldn't have connected")
|
|
|
|
def test_postgresql_opportunistically(self):
|
|
self._test_postgresql_opportunistically()
|
|
|
|
def test_postgresql_connect_fail(self):
|
|
"""
|
|
Test that we can trigger a postgres connection failure and we fail
|
|
gracefully to ensure we don't break people without postgres
|
|
"""
|
|
if _is_backend_avail('postgres', "openstack_cifail", self.PASSWD,
|
|
self.DATABASE):
|
|
self.fail("Shouldn't have connected")
|
|
|
|
|
|
class BaseMigrationTestCase(test.TestCase):
|
|
"""Base class fort testing migrations and migration utils."""
|
|
USER = None
|
|
PASSWD = None
|
|
DATABASE = None
|
|
|
|
def __init__(self, *args, **kwargs):
|
|
super(BaseMigrationTestCase, self).__init__(*args, **kwargs)
|
|
|
|
self.DEFAULT_CONFIG_FILE = os.path.join(os.path.dirname(__file__),
|
|
'test_migrations.conf')
|
|
# Test machines can set the NOVA_TEST_MIGRATIONS_CONF variable
|
|
# to override the location of the config file for migration testing
|
|
self.CONFIG_FILE_PATH = os.environ.get('NOVA_TEST_MIGRATIONS_CONF',
|
|
self.DEFAULT_CONFIG_FILE)
|
|
self.MIGRATE_FILE = nova.db.sqlalchemy.migrate_repo.__file__
|
|
self.REPOSITORY = repository.Repository(
|
|
os.path.abspath(os.path.dirname(self.MIGRATE_FILE)))
|
|
self.INIT_VERSION = 0
|
|
|
|
self.snake_walk = False
|
|
self.test_databases = {}
|
|
self.migration = None
|
|
self.migration_api = None
|
|
|
|
def setUp(self):
|
|
super(BaseMigrationTestCase, self).setUp()
|
|
|
|
# Load test databases from the config file. Only do this
|
|
# once. No need to re-run this on each test...
|
|
LOG.debug('config_path is %s' % self.CONFIG_FILE_PATH)
|
|
if os.path.exists(self.CONFIG_FILE_PATH):
|
|
cp = ConfigParser.RawConfigParser()
|
|
try:
|
|
cp.read(self.CONFIG_FILE_PATH)
|
|
defaults = cp.defaults()
|
|
for key, value in defaults.items():
|
|
self.test_databases[key] = value
|
|
self.snake_walk = cp.getboolean('walk_style', 'snake_walk')
|
|
except ConfigParser.ParsingError as e:
|
|
self.fail("Failed to read test_migrations.conf config "
|
|
"file. Got error: %s" % e)
|
|
else:
|
|
self.fail("Failed to find test_migrations.conf config "
|
|
"file.")
|
|
|
|
self.engines = {}
|
|
for key, value in self.test_databases.items():
|
|
self.engines[key] = sqlalchemy.create_engine(value)
|
|
|
|
# We start each test case with a completely blank slate.
|
|
self._reset_databases()
|
|
|
|
def tearDown(self):
|
|
# We destroy the test data store between each test case,
|
|
# and recreate it, which ensures that we have no side-effects
|
|
# from the tests
|
|
self._reset_databases()
|
|
super(BaseMigrationTestCase, self).tearDown()
|
|
|
|
def execute_cmd(self, cmd=None):
|
|
status, output = commands.getstatusoutput(cmd)
|
|
LOG.debug(output)
|
|
self.assertEqual(0, status,
|
|
"Failed to run: %s\n%s" % (cmd, output))
|
|
|
|
@utils.synchronized('pgadmin', external=True)
|
|
def _reset_pg(self, conn_pieces):
|
|
(user, password, database, host) = \
|
|
get_pgsql_connection_info(conn_pieces)
|
|
os.environ['PGPASSWORD'] = password
|
|
os.environ['PGUSER'] = user
|
|
# note(boris-42): We must create and drop database, we can't
|
|
# drop database which we have connected to, so for such
|
|
# operations there is a special database template1.
|
|
sqlcmd = ("psql -w -U %(user)s -h %(host)s -c"
|
|
" '%(sql)s' -d template1")
|
|
sqldict = {'user': user, 'host': host}
|
|
|
|
sqldict['sql'] = ("drop database if exists %s;") % database
|
|
droptable = sqlcmd % sqldict
|
|
self.execute_cmd(droptable)
|
|
|
|
sqldict['sql'] = ("create database %s;") % database
|
|
createtable = sqlcmd % sqldict
|
|
self.execute_cmd(createtable)
|
|
|
|
os.unsetenv('PGPASSWORD')
|
|
os.unsetenv('PGUSER')
|
|
|
|
def _reset_databases(self):
|
|
for key, engine in self.engines.items():
|
|
conn_string = self.test_databases[key]
|
|
conn_pieces = urlparse.urlparse(conn_string)
|
|
engine.dispose()
|
|
if conn_string.startswith('sqlite'):
|
|
# We can just delete the SQLite database, which is
|
|
# the easiest and cleanest solution
|
|
db_path = conn_pieces.path.strip('/')
|
|
if os.path.exists(db_path):
|
|
os.unlink(db_path)
|
|
# No need to recreate the SQLite DB. SQLite will
|
|
# create it for us if it's not there...
|
|
elif conn_string.startswith('mysql'):
|
|
# We can execute the MySQL client to destroy and re-create
|
|
# the MYSQL database, which is easier and less error-prone
|
|
# than using SQLAlchemy to do this via MetaData...trust me.
|
|
(user, password, database, host) = \
|
|
get_mysql_connection_info(conn_pieces)
|
|
sql = ("drop database if exists %(database)s; "
|
|
"create database %(database)s;"
|
|
% {'database': database})
|
|
cmd = ("mysql -u \"%(user)s\" %(password)s -h %(host)s "
|
|
"-e \"%(sql)s\"" % {'user': user,
|
|
'password': password, 'host': host, 'sql': sql})
|
|
self.execute_cmd(cmd)
|
|
elif conn_string.startswith('postgresql'):
|
|
self._reset_pg(conn_pieces)
|
|
|
|
def _test_mysql_opportunistically(self):
|
|
# Test that table creation on mysql only builds InnoDB tables
|
|
if not _have_mysql(self.USER, self.PASSWD, self.DATABASE):
|
|
self.skipTest("mysql not available")
|
|
# add this to the global lists to make reset work with it, it's removed
|
|
# automatically in tearDown so no need to clean it up here.
|
|
connect_string = _get_connect_string("mysql", self.USER, self.PASSWD,
|
|
self.DATABASE)
|
|
(user, password, database, host) = \
|
|
get_mysql_connection_info(urlparse.urlparse(connect_string))
|
|
engine = sqlalchemy.create_engine(connect_string)
|
|
self.engines[database] = engine
|
|
self.test_databases[database] = connect_string
|
|
|
|
# build a fully populated mysql database with all the tables
|
|
self._reset_databases()
|
|
self._walk_versions(engine, False, False)
|
|
|
|
connection = engine.connect()
|
|
# sanity check
|
|
total = connection.execute("SELECT count(*) "
|
|
"from information_schema.TABLES "
|
|
"where TABLE_SCHEMA='%(database)s'" %
|
|
{'database': database})
|
|
self.assertTrue(total.scalar() > 0, "No tables found. Wrong schema?")
|
|
|
|
noninnodb = connection.execute("SELECT count(*) "
|
|
"from information_schema.TABLES "
|
|
"where TABLE_SCHEMA='%(database)s' "
|
|
"and ENGINE!='InnoDB' "
|
|
"and TABLE_NAME!='migrate_version'" %
|
|
{'database': database})
|
|
count = noninnodb.scalar()
|
|
self.assertEqual(count, 0, "%d non InnoDB tables created" % count)
|
|
connection.close()
|
|
|
|
def _test_postgresql_opportunistically(self):
|
|
# Test postgresql database migration walk
|
|
if not _have_postgresql(self.USER, self.PASSWD, self.DATABASE):
|
|
self.skipTest("postgresql not available")
|
|
# add this to the global lists to make reset work with it, it's removed
|
|
# automatically in tearDown so no need to clean it up here.
|
|
connect_string = _get_connect_string("postgres", self.USER,
|
|
self.PASSWD, self.DATABASE)
|
|
engine = sqlalchemy.create_engine(connect_string)
|
|
(user, password, database, host) = \
|
|
get_mysql_connection_info(urlparse.urlparse(connect_string))
|
|
self.engines[database] = engine
|
|
self.test_databases[database] = connect_string
|
|
|
|
# build a fully populated postgresql database with all the tables
|
|
self._reset_databases()
|
|
self._walk_versions(engine, False, False)
|
|
|
|
def _walk_versions(self, engine=None, snake_walk=False, downgrade=True):
|
|
# Determine latest version script from the repo, then
|
|
# upgrade from 1 through to the latest, with no data
|
|
# in the databases. This just checks that the schema itself
|
|
# upgrades successfully.
|
|
|
|
# Place the database under version control
|
|
self.migration_api.version_control(engine,
|
|
self.REPOSITORY,
|
|
self.INIT_VERSION)
|
|
self.assertEqual(self.INIT_VERSION,
|
|
self.migration_api.db_version(engine,
|
|
self.REPOSITORY))
|
|
|
|
LOG.debug('latest version is %s' % self.REPOSITORY.latest)
|
|
versions = range(self.INIT_VERSION + 1, self.REPOSITORY.latest + 1)
|
|
|
|
for version in versions:
|
|
# upgrade -> downgrade -> upgrade
|
|
self._migrate_up(engine, version, with_data=True)
|
|
if snake_walk:
|
|
downgraded = self._migrate_down(
|
|
engine, version - 1, with_data=True)
|
|
if downgraded:
|
|
self._migrate_up(engine, version)
|
|
|
|
if downgrade:
|
|
# Now walk it back down to 0 from the latest, testing
|
|
# the downgrade paths.
|
|
for version in reversed(versions):
|
|
# downgrade -> upgrade -> downgrade
|
|
downgraded = self._migrate_down(engine, version - 1)
|
|
|
|
if snake_walk and downgraded:
|
|
self._migrate_up(engine, version)
|
|
self._migrate_down(engine, version - 1)
|
|
|
|
def _migrate_down(self, engine, version, with_data=False):
|
|
try:
|
|
self.migration_api.downgrade(engine, self.REPOSITORY, version)
|
|
except NotImplementedError:
|
|
# NOTE(sirp): some migrations, namely release-level
|
|
# migrations, don't support a downgrade.
|
|
return False
|
|
|
|
self.assertEqual(version,
|
|
self.migration_api.db_version(engine,
|
|
self.REPOSITORY))
|
|
|
|
# NOTE(sirp): `version` is what we're downgrading to (i.e. the 'target'
|
|
# version). So if we have any downgrade checks, they need to be run for
|
|
# the previous (higher numbered) migration.
|
|
if with_data:
|
|
post_downgrade = getattr(
|
|
self, "_post_downgrade_%03d" % (version + 1), None)
|
|
if post_downgrade:
|
|
post_downgrade(engine)
|
|
|
|
return True
|
|
|
|
def _migrate_up(self, engine, version, with_data=False):
|
|
"""migrate up to a new version of the db.
|
|
|
|
We allow for data insertion and post checks at every
|
|
migration version with special _pre_upgrade_### and
|
|
_check_### functions in the main test.
|
|
"""
|
|
# NOTE(sdague): try block is here because it's impossible to debug
|
|
# where a failed data migration happens otherwise
|
|
try:
|
|
if with_data:
|
|
data = None
|
|
pre_upgrade = getattr(
|
|
self, "_pre_upgrade_%03d" % version, None)
|
|
if pre_upgrade:
|
|
data = pre_upgrade(engine)
|
|
|
|
self.migration_api.upgrade(engine, self.REPOSITORY, version)
|
|
self.assertEqual(version,
|
|
self.migration_api.db_version(engine,
|
|
self.REPOSITORY))
|
|
if with_data:
|
|
check = getattr(self, "_check_%03d" % version, None)
|
|
if check:
|
|
check(engine, data)
|
|
except Exception:
|
|
LOG.error("Failed to migrate to version %s on engine %s" %
|
|
(version, engine))
|
|
raise
|
|
|
|
|
|
class TestNovaMigrations(BaseMigrationTestCase, CommonTestsMixIn):
|
|
"""Test sqlalchemy-migrate migrations."""
|
|
USER = "openstack_citest"
|
|
PASSWD = "openstack_citest"
|
|
DATABASE = "openstack_citest"
|
|
|
|
def __init__(self, *args, **kwargs):
|
|
super(TestNovaMigrations, self).__init__(*args, **kwargs)
|
|
|
|
self.DEFAULT_CONFIG_FILE = os.path.join(os.path.dirname(__file__),
|
|
'test_migrations.conf')
|
|
# Test machines can set the NOVA_TEST_MIGRATIONS_CONF variable
|
|
# to override the location of the config file for migration testing
|
|
self.CONFIG_FILE_PATH = os.environ.get('NOVA_TEST_MIGRATIONS_CONF',
|
|
self.DEFAULT_CONFIG_FILE)
|
|
self.MIGRATE_FILE = nova.db.sqlalchemy.migrate_repo.__file__
|
|
self.REPOSITORY = repository.Repository(
|
|
os.path.abspath(os.path.dirname(self.MIGRATE_FILE)))
|
|
|
|
def setUp(self):
|
|
super(TestNovaMigrations, self).setUp()
|
|
|
|
if self.migration is None:
|
|
self.migration = __import__('nova.db.migration',
|
|
globals(), locals(), ['INIT_VERSION'], -1)
|
|
self.INIT_VERSION = self.migration.INIT_VERSION
|
|
if self.migration_api is None:
|
|
temp = __import__('nova.db.sqlalchemy.migration',
|
|
globals(), locals(), ['versioning_api'], -1)
|
|
self.migration_api = temp.versioning_api
|
|
|
|
def assertColumnExists(self, engine, table, column):
|
|
t = db_utils.get_table(engine, table)
|
|
self.assertIn(column, t.c)
|
|
|
|
def assertColumnNotExists(self, engine, table, column):
|
|
t = db_utils.get_table(engine, table)
|
|
self.assertNotIn(column, t.c)
|
|
|
|
def assertIndexExists(self, engine, table, index):
|
|
t = db_utils.get_table(engine, table)
|
|
index_names = [idx.name for idx in t.indexes]
|
|
self.assertIn(index, index_names)
|
|
|
|
def assertIndexMembers(self, engine, table, index, members):
|
|
self.assertIndexExists(engine, table, index)
|
|
|
|
t = db_utils.get_table(engine, table)
|
|
index_columns = None
|
|
for idx in t.indexes:
|
|
if idx.name == index:
|
|
index_columns = idx.columns.keys()
|
|
break
|
|
|
|
self.assertEqual(sorted(members), sorted(index_columns))
|
|
|
|
def _pre_upgrade_134(self, engine):
|
|
now = timeutils.utcnow()
|
|
data = [{
|
|
'id': 1,
|
|
'uuid': '1d739808-d7ec-4944-b252-f8363e119755',
|
|
'mac': '00:00:00:00:00:01',
|
|
'start_period': now,
|
|
'last_refreshed': now + datetime.timedelta(seconds=10),
|
|
'bw_in': 100000,
|
|
'bw_out': 200000,
|
|
}, {
|
|
'id': 2,
|
|
'uuid': '1d739808-d7ec-4944-b252-f8363e119756',
|
|
'mac': '2a:f2:48:31:c1:60',
|
|
'start_period': now,
|
|
'last_refreshed': now + datetime.timedelta(seconds=20),
|
|
'bw_in': 1000000000,
|
|
'bw_out': 200000000,
|
|
}, {
|
|
'id': 3,
|
|
# This is intended to be the same as above.
|
|
'uuid': '1d739808-d7ec-4944-b252-f8363e119756',
|
|
'mac': '00:00:00:00:00:02',
|
|
'start_period': now,
|
|
'last_refreshed': now + datetime.timedelta(seconds=30),
|
|
'bw_in': 0,
|
|
'bw_out': 0,
|
|
}]
|
|
|
|
bw_usage_cache = db_utils.get_table(engine, 'bw_usage_cache')
|
|
engine.execute(bw_usage_cache.insert(), data)
|
|
return data
|
|
|
|
def _check_134(self, engine, data):
|
|
bw_usage_cache = db_utils.get_table(engine, 'bw_usage_cache')
|
|
|
|
# Checks if both columns have been successfuly created.
|
|
self.assertIn('last_ctr_in', bw_usage_cache.c)
|
|
self.assertIn('last_ctr_out', bw_usage_cache.c)
|
|
|
|
# Checks if all rows have been inserted.
|
|
bw_items = bw_usage_cache.select().execute().fetchall()
|
|
self.assertEqual(len(bw_items), 3)
|
|
|
|
bw = bw_usage_cache.select(
|
|
bw_usage_cache.c.id == 1).execute().first()
|
|
|
|
# New columns have 'NULL' as default value.
|
|
self.assertEqual(bw['last_ctr_in'], None)
|
|
self.assertEqual(bw['last_ctr_out'], None)
|
|
|
|
self.assertEqual(data[0]['mac'], bw['mac'])
|
|
|
|
# migration 141, update migrations instance uuid
|
|
def _pre_upgrade_141(self, engine):
|
|
data = {
|
|
'instance_uuid': str(uuid.uuid4())
|
|
}
|
|
migrations = db_utils.get_table(engine, 'migrations')
|
|
engine.execute(migrations.insert(), data)
|
|
result = migrations.insert().values(data).execute()
|
|
data['id'] = result.inserted_primary_key[0]
|
|
return data
|
|
|
|
def _check_141(self, engine, data):
|
|
migrations = db_utils.get_table(engine, 'migrations')
|
|
row = migrations.select(
|
|
migrations.c.id == data['id']).execute().first()
|
|
# Check that change to String(36) went alright
|
|
self.assertEqual(data['instance_uuid'], row['instance_uuid'])
|
|
|
|
# migration 146, availability zone transition
|
|
def _pre_upgrade_146(self, engine):
|
|
data = {
|
|
'availability_zone': 'custom_az',
|
|
'name': 'name',
|
|
}
|
|
|
|
aggregates = db_utils.get_table(engine, 'aggregates')
|
|
result = aggregates.insert().values(data).execute()
|
|
# NOTE(sdague) it's important you don't insert keys by value in
|
|
# postgresql, because its autoincrement counter won't get updated
|
|
data['id'] = result.inserted_primary_key[0]
|
|
return data
|
|
|
|
def _check_146(self, engine, data):
|
|
aggregate_md = db_utils.get_table(engine, 'aggregate_metadata')
|
|
md = aggregate_md.select(
|
|
aggregate_md.c.aggregate_id == data['id']).execute().first()
|
|
self.assertEqual(data['availability_zone'], md['value'])
|
|
|
|
def _post_downgrade_146(self, engine):
|
|
# Downgrade should delete availability_zone aggregate_metadata entries
|
|
aggregate_md = db_utils.get_table(engine, 'aggregate_metadata')
|
|
num_azs = aggregate_md.count().where(
|
|
aggregate_md.c.key == 'availability_zone').execute().scalar()
|
|
self.assertEqual(0, num_azs)
|
|
|
|
def _upgrade_147_test_data(self):
|
|
az = 'test_zone'
|
|
host1 = 'compute-host1'
|
|
host2 = 'compute-host2'
|
|
# start at id == 2 because we already inserted one
|
|
data = [
|
|
{'id': 1, 'host': host1,
|
|
'binary': 'nova-compute', 'topic': 'compute',
|
|
'report_count': 0, 'availability_zone': az},
|
|
{'id': 2, 'host': 'sched-host',
|
|
'binary': 'nova-scheduler', 'topic': 'scheduler',
|
|
'report_count': 0, 'availability_zone': 'ignore_me'},
|
|
{'id': 3, 'host': host2,
|
|
'binary': 'nova-compute', 'topic': 'compute',
|
|
'report_count': 0, 'availability_zone': az},
|
|
]
|
|
|
|
return data
|
|
|
|
# migration 147, availability zone transition for services
|
|
def _pre_upgrade_147(self, engine):
|
|
data = self._upgrade_147_test_data()
|
|
services = db_utils.get_table(engine, 'services')
|
|
engine.execute(services.insert(), data)
|
|
self._pre_upgrade_147_no_duplicate_aggregate_hosts(engine)
|
|
return data
|
|
|
|
def _pre_upgrade_147_no_duplicate_aggregate_hosts(self, engine):
|
|
engine.execute(
|
|
db_utils.get_table(engine, 'aggregate_metadata').insert(),
|
|
[{'aggregate_id': 1,
|
|
'key': 'availability_zone',
|
|
'value': 'custom_az'}])
|
|
|
|
engine.execute(db_utils.get_table(engine, 'aggregate_hosts').insert(),
|
|
[{'aggregate_id': 1,
|
|
'host': 'compute-host3'}])
|
|
|
|
engine.execute(db_utils.get_table(engine, 'services').insert(),
|
|
[{'id': 99, 'host': 'compute-host3',
|
|
'binary': 'nova-compute', 'topic': 'compute',
|
|
'report_count': 0, 'availability_zone': 'custom_az'}])
|
|
|
|
def _check_147(self, engine, data):
|
|
aggregate_md = db_utils.get_table(engine, 'aggregate_metadata')
|
|
aggregate_hosts = db_utils.get_table(engine, 'aggregate_hosts')
|
|
# NOTE(sdague): hard coded to id == 2, because we added to
|
|
# aggregate_metadata previously
|
|
for item in data:
|
|
md = aggregate_md.select(
|
|
aggregate_md.c.aggregate_id == 2).execute().first()
|
|
if item['binary'] == "nova-compute":
|
|
self.assertEqual(item['availability_zone'], md['value'])
|
|
|
|
host = aggregate_hosts.select(
|
|
aggregate_hosts.c.aggregate_id == 2
|
|
).execute().first()
|
|
self.assertEqual(host['host'], data[0]['host'])
|
|
|
|
# NOTE(sdague): id 3 is just non-existent
|
|
host = aggregate_hosts.select(
|
|
aggregate_hosts.c.aggregate_id == 3
|
|
).execute().first()
|
|
self.assertEqual(host, None)
|
|
|
|
self._check_147_no_duplicate_aggregate_hosts(engine, data)
|
|
|
|
def _check_147_no_duplicate_aggregate_hosts(self, engine, data):
|
|
aggregate_hosts = db_utils.get_table(engine, 'aggregate_hosts')
|
|
agg1_hosts = [h['host'] for h in aggregate_hosts.select(
|
|
aggregate_hosts.c.aggregate_id == 1
|
|
).execute().fetchall()]
|
|
self.assertEqual(['compute-host3'], agg1_hosts)
|
|
|
|
def _post_downgrade_147(self, engine):
|
|
# Test that availability_zone is back on the services table.
|
|
services = db_utils.get_table(engine, 'services')
|
|
record_list = list(services.select().execute())
|
|
test_data = self._upgrade_147_test_data()
|
|
|
|
availability_zones = [x['availability_zone'] for x in test_data]
|
|
# Append the default availability_zone
|
|
availability_zones.append('nova')
|
|
|
|
for row in record_list:
|
|
self.assertIn(row['availability_zone'], availability_zones)
|
|
|
|
# migration 149, changes IPAddr storage format
|
|
def _pre_upgrade_149(self, engine):
|
|
provider_fw_rules = db_utils.get_table(engine, 'provider_fw_rules')
|
|
console_pools = db_utils.get_table(engine, 'console_pools')
|
|
data = {
|
|
'provider_fw_rules':
|
|
[
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': str(netaddr.IPNetwork(
|
|
"127.0.0.1/30"))},
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': str(netaddr.IPNetwork(
|
|
"128.128.128.128/16"))},
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': str(netaddr.IPNetwork(
|
|
"128.128.128.128/32"))},
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': str(netaddr.IPNetwork(
|
|
"2001:db8::1:2/48"))},
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': str(netaddr.IPNetwork(
|
|
"::1/64"))},
|
|
{'protocol': 'tcp', 'from_port': 1234, 'to_port': 1234,
|
|
'cidr': str(netaddr.IPNetwork(
|
|
"0000:0000:0000:2013:0000:6535:abcd:ef11/64"))},
|
|
{'protocol': 'tcp', 'from_port': 1234, 'to_port': 1234,
|
|
'cidr': str(netaddr.IPNetwork(
|
|
"0000:1020:0000:2013:0000:6535:abcd:ef11/128"))},
|
|
],
|
|
'console_pools':
|
|
[
|
|
{'address': '10.10.10.10'},
|
|
{'address': '128.100.100.100'},
|
|
{'address': '2002:2002:2002:2002:2002:2002:2002:2002'},
|
|
{'address': '::1'},
|
|
{'address': '0000:0000:0000:2013:0000:6535:abcd:ef11'}
|
|
]
|
|
}
|
|
|
|
engine.execute(provider_fw_rules.insert(), data['provider_fw_rules'])
|
|
|
|
for pool in data['console_pools']:
|
|
engine.execute(console_pools.insert(), pool)
|
|
|
|
return data
|
|
|
|
def _check_149(self, engine, data):
|
|
provider_fw_rules = db_utils.get_table(engine, 'provider_fw_rules')
|
|
result = provider_fw_rules.select().execute()
|
|
|
|
iplist = map(lambda x: str(netaddr.IPNetwork(x['cidr'])),
|
|
data['provider_fw_rules'])
|
|
|
|
for row in result:
|
|
self.assertIn(str(netaddr.IPNetwork(row['cidr'])), iplist)
|
|
|
|
console_pools = db_utils.get_table(engine, 'console_pools')
|
|
result = console_pools.select().execute()
|
|
|
|
iplist = map(lambda x: str(netaddr.IPAddress(x['address'])),
|
|
data['console_pools'])
|
|
|
|
for row in result:
|
|
self.assertIn(str(netaddr.IPAddress(row['address'])), iplist)
|
|
|
|
# migration 151 - changes period_beginning and period_ending to DateTime
|
|
def _pre_upgrade_151(self, engine):
|
|
task_log = db_utils.get_table(engine, 'task_log')
|
|
data = {
|
|
'task_name': 'The name of the task',
|
|
'state': 'The state of the task',
|
|
'host': 'compute-host1',
|
|
'period_beginning': str(datetime.datetime(2013, 2, 11)),
|
|
'period_ending': str(datetime.datetime(2013, 2, 12)),
|
|
'message': 'The task_log message',
|
|
}
|
|
result = task_log.insert().values(data).execute()
|
|
data['id'] = result.inserted_primary_key[0]
|
|
return data
|
|
|
|
def _check_151(self, engine, data):
|
|
task_log = db_utils.get_table(engine, 'task_log')
|
|
row = task_log.select(task_log.c.id == data['id']).execute().first()
|
|
self.assertTrue(isinstance(row['period_beginning'],
|
|
datetime.datetime))
|
|
self.assertTrue(isinstance(row['period_ending'],
|
|
datetime.datetime))
|
|
self.assertEqual(
|
|
data['period_beginning'], str(row['period_beginning']))
|
|
self.assertEqual(data['period_ending'], str(row['period_ending']))
|
|
|
|
# migration 152 - convert deleted from boolean to int
|
|
def _pre_upgrade_152(self, engine):
|
|
host1 = 'compute-host1'
|
|
host2 = 'compute-host2'
|
|
# NOTE(sdague): start at #4 because services data already in table
|
|
# from 147
|
|
services_data = [
|
|
{'id': 4, 'host': host1, 'binary': 'nova-compute',
|
|
'report_count': 0, 'topic': 'compute', 'deleted': False},
|
|
{'id': 5, 'host': host1, 'binary': 'nova-compute',
|
|
'report_count': 0, 'topic': 'compute', 'deleted': True}
|
|
]
|
|
volumes_data = [
|
|
{'id': 'first', 'host': host1, 'deleted': False},
|
|
{'id': 'second', 'host': host2, 'deleted': True}
|
|
]
|
|
|
|
services = db_utils.get_table(engine, 'services')
|
|
engine.execute(services.insert(), services_data)
|
|
|
|
volumes = db_utils.get_table(engine, 'volumes')
|
|
engine.execute(volumes.insert(), volumes_data)
|
|
return dict(services=services_data, volumes=volumes_data)
|
|
|
|
def _check_152(self, engine, data):
|
|
services = db_utils.get_table(engine, 'services')
|
|
service = services.select(services.c.id == 4).execute().first()
|
|
self.assertEqual(0, service.deleted)
|
|
service = services.select(services.c.id == 5).execute().first()
|
|
self.assertEqual(service.id, service.deleted)
|
|
|
|
volumes = db_utils.get_table(engine, 'volumes')
|
|
volume = volumes.select(volumes.c.id == "first").execute().first()
|
|
self.assertEqual("", volume.deleted)
|
|
volume = volumes.select(volumes.c.id == "second").execute().first()
|
|
self.assertEqual(volume.id, volume.deleted)
|
|
|
|
# migration 153, copy flavor information into system_metadata
|
|
def _pre_upgrade_153(self, engine):
|
|
fake_types = [
|
|
dict(id=10, name='type1', memory_mb=128, vcpus=1,
|
|
root_gb=10, ephemeral_gb=0, flavorid="1", swap=0,
|
|
rxtx_factor=1.0, vcpu_weight=1, disabled=False,
|
|
is_public=True),
|
|
dict(id=11, name='type2', memory_mb=512, vcpus=1,
|
|
root_gb=10, ephemeral_gb=5, flavorid="2", swap=0,
|
|
rxtx_factor=1.5, vcpu_weight=2, disabled=False,
|
|
is_public=True),
|
|
dict(id=12, name='type3', memory_mb=128, vcpus=1,
|
|
root_gb=10, ephemeral_gb=0, flavorid="3", swap=0,
|
|
rxtx_factor=1.0, vcpu_weight=1, disabled=False,
|
|
is_public=False),
|
|
dict(id=13, name='type4', memory_mb=128, vcpus=1,
|
|
root_gb=10, ephemeral_gb=0, flavorid="4", swap=0,
|
|
rxtx_factor=1.0, vcpu_weight=None, disabled=True,
|
|
is_public=True),
|
|
dict(id=14, name='type5', memory_mb=128, vcpus=1,
|
|
root_gb=10, ephemeral_gb=0, flavorid="5", swap=0,
|
|
rxtx_factor=1.0, vcpu_weight=1, disabled=True,
|
|
is_public=False),
|
|
]
|
|
|
|
fake_instances = [
|
|
dict(uuid='m153-uuid1', instance_type_id=10, deleted=0),
|
|
dict(uuid='m153-uuid2', instance_type_id=11, deleted=0),
|
|
dict(uuid='m153-uuid3', instance_type_id=12, deleted=0),
|
|
dict(uuid='m153-uuid4', instance_type_id=13, deleted=0),
|
|
# NOTE(danms): no use of type5
|
|
]
|
|
|
|
instances = db_utils.get_table(engine, 'instances')
|
|
instance_types = db_utils.get_table(engine, 'instance_types')
|
|
engine.execute(instance_types.insert(), fake_types)
|
|
engine.execute(instances.insert(), fake_instances)
|
|
|
|
return fake_types, fake_instances
|
|
|
|
def _check_153(self, engine, data):
|
|
fake_types, fake_instances = data
|
|
# NOTE(danms): Fetch all the tables and data from scratch after change
|
|
sys_meta = db_utils.get_table(engine, 'instance_system_metadata')
|
|
|
|
# Collect all system metadata, indexed by instance_uuid
|
|
metadata = collections.defaultdict(dict)
|
|
for values in sys_meta.select().execute():
|
|
metadata[values['instance_uuid']][values['key']] = values['value']
|
|
|
|
# Taken from nova/compute/api.py
|
|
instance_type_props = ['id', 'name', 'memory_mb', 'vcpus',
|
|
'root_gb', 'ephemeral_gb', 'flavorid',
|
|
'swap', 'rxtx_factor', 'vcpu_weight']
|
|
|
|
for instance in fake_instances:
|
|
inst_sys_meta = metadata[instance['uuid']]
|
|
inst_type = fake_types[instance['instance_type_id'] - 10]
|
|
for prop in instance_type_props:
|
|
prop_name = 'instance_type_%s' % prop
|
|
self.assertIn(prop_name, inst_sys_meta)
|
|
if prop == "vcpu_weight":
|
|
# NOTE(danms) vcpu_weight can be NULL
|
|
self.assertEqual(inst_sys_meta[prop_name],
|
|
inst_type[prop] and str(inst_type[prop])
|
|
or None)
|
|
else:
|
|
self.assertEqual(str(inst_sys_meta[prop_name]),
|
|
str(inst_type[prop]))
|
|
|
|
# migration 154, add shadow tables for deleted data
|
|
# There are 53 shadow tables but we only test one
|
|
# There are additional tests in test_db_api.py
|
|
def _pre_upgrade_154(self, engine):
|
|
meta = sqlalchemy.schema.MetaData()
|
|
meta.reflect(engine)
|
|
table_names = meta.tables.keys()
|
|
for table_name in table_names:
|
|
self.assertFalse(table_name.startswith("_shadow"))
|
|
|
|
def _check_154(self, engine, data):
|
|
meta = sqlalchemy.schema.MetaData()
|
|
meta.reflect(engine)
|
|
table_names = set(meta.tables.keys())
|
|
for table_name in table_names:
|
|
if table_name.startswith(db._SHADOW_TABLE_PREFIX):
|
|
shadow_name = table_name
|
|
base_name = table_name.replace(db._SHADOW_TABLE_PREFIX, "")
|
|
self.assertIn(base_name, table_names)
|
|
else:
|
|
base_name = table_name
|
|
shadow_name = db._SHADOW_TABLE_PREFIX + table_name
|
|
self.assertIn(shadow_name, table_names)
|
|
shadow_table = db_utils.get_table(engine, shadow_name)
|
|
base_table = db_utils.get_table(engine, base_name)
|
|
base_columns = []
|
|
shadow_columns = []
|
|
for column in base_table.columns:
|
|
base_columns.append(column)
|
|
for column in shadow_table.columns:
|
|
shadow_columns.append(column)
|
|
for ii, base_column in enumerate(base_columns):
|
|
shadow_column = shadow_columns[ii]
|
|
self.assertEqual(base_column.name, shadow_column.name)
|
|
# NullType needs a special case. We end up with NullType on sqlite
|
|
# where bigint is not defined.
|
|
if isinstance(base_column.type, sqlalchemy.types.NullType):
|
|
self.assertTrue(isinstance(shadow_column.type,
|
|
sqlalchemy.types.NullType))
|
|
else:
|
|
# Identical types do not test equal because sqlalchemy does not
|
|
# override __eq__, but if we stringify them then they do.
|
|
self.assertEqual(str(base_column.type),
|
|
str(shadow_column.type))
|
|
|
|
# migration 156 - introduce CIDR type
|
|
def _pre_upgrade_156(self, engine):
|
|
# assume the same data as from 149
|
|
data = {
|
|
'provider_fw_rules':
|
|
[
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': str(netaddr.IPNetwork(
|
|
"127.0.0.1/30"))},
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': str(netaddr.IPNetwork(
|
|
"128.128.128.128/16"))},
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': str(netaddr.IPNetwork(
|
|
"128.128.128.128/32"))},
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': str(netaddr.IPNetwork(
|
|
"2001:db8::1:2/48"))},
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': str(netaddr.IPNetwork(
|
|
"::1/64"))},
|
|
{'protocol': 'tcp', 'from_port': 1234, 'to_port': 1234,
|
|
'cidr': str(netaddr.IPNetwork(
|
|
"0000:0000:0000:2013:0000:6535:abcd:ef11/64"))},
|
|
{'protocol': 'tcp', 'from_port': 1234, 'to_port': 1234,
|
|
'cidr': str(netaddr.IPNetwork(
|
|
"0000:1020:0000:2013:0000:6535:abcd:ef11/128"))},
|
|
],
|
|
'console_pools':
|
|
[
|
|
{'address': '10.10.10.10'},
|
|
{'address': '128.100.100.100'},
|
|
{'address': '2002:2002:2002:2002:2002:2002:2002:2002'},
|
|
{'address': '::1'},
|
|
{'address': '0000:0000:0000:2013:0000:6535:abcd:ef11'}
|
|
]
|
|
}
|
|
return data
|
|
|
|
def _check_156(self, engine, data):
|
|
# recheck the 149 data
|
|
self._check_149(engine, data)
|
|
|
|
def _pre_upgrade_158(self, engine):
|
|
networks = db_utils.get_table(engine, 'networks')
|
|
data = [
|
|
{'vlan': 1, 'deleted': 0},
|
|
{'vlan': 1, 'deleted': 0},
|
|
{'vlan': 1, 'deleted': 0},
|
|
]
|
|
|
|
for item in data:
|
|
networks.insert().values(item).execute()
|
|
return data
|
|
|
|
def _check_158(self, engine, data):
|
|
networks = db_utils.get_table(engine, 'networks')
|
|
rows = networks.select().\
|
|
where(networks.c.deleted != networks.c.id).\
|
|
execute().\
|
|
fetchall()
|
|
self.assertEqual(len(rows), 1)
|
|
|
|
def _pre_upgrade_159(self, engine):
|
|
data = {
|
|
'provider_fw_rules':
|
|
[
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': "127.0.0.1/30"},
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': "128.128.128.128/16"},
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': "128.128.128.128/32"},
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': "2001:db8::1:2/48"},
|
|
{'protocol': 'tcp', 'from_port': 1234,
|
|
'to_port': 1234, 'cidr': "::1/64"},
|
|
{'protocol': 'tcp', 'from_port': 1234, 'to_port': 1234,
|
|
'cidr': "0000:0000:0000:2013:0000:6535:abcd:ef11/64"},
|
|
{'protocol': 'tcp', 'from_port': 1234, 'to_port': 1234,
|
|
'cidr': "0000:1020:0000:2013:0000:6535:abcd:ef11/128"},
|
|
],
|
|
'console_pools':
|
|
[
|
|
{'address': '10.10.10.10'},
|
|
{'address': '128.100.100.100'},
|
|
{'address': '2002:2002:2002:2002:2002:2002:2002:2002'},
|
|
{'address': '::1'},
|
|
{'address': '0000:0000:0000:2013:0000:6535:abcd:ef11'}
|
|
]
|
|
}
|
|
return data
|
|
|
|
# migration 159 - revert ip column size
|
|
def _check_159(self, engine, data):
|
|
dialect = engine.url.get_dialect()
|
|
# NOTE(maurosr): check if column length is 39 again (it currently makes
|
|
# sense only for mysql)
|
|
if dialect not in [postgresql.dialect, sqlite.dialect]:
|
|
console_pools = db_utils.get_table(engine, 'console_pools')
|
|
self.assertEqual(console_pools.columns['address'].type.length, 39)
|
|
# recheck the 149 data
|
|
self._check_149(engine, data)
|
|
|
|
def _post_downgrade_159(self, engine):
|
|
dialect = engine.url.get_dialect()
|
|
# NOTE(maurosr): check if column length is 43 again (it currently makes
|
|
# sense only for mysql)
|
|
if dialect not in [postgresql.dialect, sqlite.dialect]:
|
|
console_pools = db_utils.get_table(engine, 'console_pools')
|
|
self.assertEqual(console_pools.columns['address'].type.length, 43)
|
|
|
|
# migration 160, fix system_metadata NULL deleted entries to be 0
|
|
def _pre_upgrade_160(self, engine):
|
|
fake_instances = [
|
|
dict(uuid='m160-uuid1'),
|
|
dict(uuid='m160-uuid2'),
|
|
dict(uuid='m160-uuid3'),
|
|
]
|
|
fake_sys_meta = [
|
|
dict(instance_uuid='m160-uuid1', key='foo', value='bar'),
|
|
dict(instance_uuid='m160-uuid2', key='foo2', value='bar2'),
|
|
dict(instance_uuid='m160-uuid3', key='foo3', value='bar3')]
|
|
|
|
instances = db_utils.get_table(engine, 'instances')
|
|
sys_meta = db_utils.get_table(engine, 'instance_system_metadata')
|
|
engine.execute(instances.insert(), fake_instances)
|
|
|
|
# Create the metadata entries
|
|
data = {}
|
|
for sm in fake_sys_meta:
|
|
result = sys_meta.insert().values(sm).execute()
|
|
sm['id'] = result.inserted_primary_key[0]
|
|
data[sm['id']] = sm
|
|
|
|
# Make sure the entries in the DB for 'deleted' are None.
|
|
our_ids = data.keys()
|
|
results = sys_meta.select().where(sys_meta.c.id.in_(our_ids)).\
|
|
execute()
|
|
results = list(results)
|
|
self.assertEqual(len(our_ids), len(results))
|
|
for result in results:
|
|
self.assertEqual(result['deleted'], None)
|
|
return data
|
|
|
|
def _check_160(self, engine, data):
|
|
our_ids = data.keys()
|
|
sys_meta = db_utils.get_table(engine, 'instance_system_metadata')
|
|
results = sys_meta.select().where(sys_meta.c.id.in_(our_ids)).\
|
|
execute()
|
|
results = list(results)
|
|
self.assertEqual(len(our_ids), len(results))
|
|
for result in results:
|
|
the_id = result['id']
|
|
# Make sure this is now 0.
|
|
self.assertEqual(result['deleted'], 0)
|
|
# Make sure nothing else changed.
|
|
for key, value in data[the_id].items():
|
|
self.assertEqual(value, result[key])
|
|
|
|
# migration 161, fix system_metadata "None" values should be NULL
|
|
def _pre_upgrade_161(self, engine):
|
|
fake_instances = [dict(uuid='m161-uuid1')]
|
|
sm_base = dict(instance_uuid='m161-uuid1', value=None)
|
|
now = timeutils.utcnow().replace(microsecond=0)
|
|
fake_sys_meta = [
|
|
# Should be fixed
|
|
dict(sm_base, key='instance_type_foo', value='None'),
|
|
dict(sm_base, key='instance_type_bar', value='88 mph'),
|
|
|
|
# Should be unaffected
|
|
dict(sm_base, key='instance_type_name', value='None'),
|
|
dict(sm_base, key='instance_type_flavorid', value='None'),
|
|
dict(sm_base, key='foo', value='None'),
|
|
dict(sm_base, key='instance_type_bat'),
|
|
dict(sm_base, key='instance_type_baz', created_at=now),
|
|
]
|
|
|
|
instances = db_utils.get_table(engine, 'instances')
|
|
sys_meta = db_utils.get_table(engine, 'instance_system_metadata')
|
|
engine.execute(instances.insert(), fake_instances)
|
|
|
|
data = {}
|
|
for sm in fake_sys_meta:
|
|
result = sys_meta.insert().values(sm).execute()
|
|
sm['id'] = result.inserted_primary_key[0]
|
|
data[sm['id']] = sm
|
|
|
|
return data
|
|
|
|
def _check_161(self, engine, data):
|
|
our_ids = data.keys()
|
|
sys_meta = db_utils.get_table(engine, 'instance_system_metadata')
|
|
results = sys_meta.select().where(sys_meta.c.id.in_(our_ids)).\
|
|
execute()
|
|
results = list(results)
|
|
self.assertEqual(len(our_ids), len(results))
|
|
for result in results:
|
|
the_id = result['id']
|
|
key = result['key']
|
|
original = data[the_id]
|
|
|
|
if key == 'instance_type_baz':
|
|
# Neither value nor created_at should have been altered
|
|
self.assertEqual(result['value'], original['value'])
|
|
self.assertEqual(result['created_at'], original['created_at'])
|
|
elif key in ['instance_type_name', 'instance_type_flavorid']:
|
|
# These should not have their values changed, but should
|
|
# have corrected created_at stamps
|
|
self.assertEqual(result['value'], original['value'])
|
|
self.assertTrue(isinstance(result['created_at'],
|
|
datetime.datetime))
|
|
elif key.startswith('instance_type'):
|
|
# Values like instance_type_% should be stamped and values
|
|
# converted from 'None' to None where appropriate
|
|
self.assertEqual(result['value'],
|
|
None if original['value'] == 'None'
|
|
else original['value'])
|
|
self.assertTrue(isinstance(result['created_at'],
|
|
datetime.datetime))
|
|
else:
|
|
# None of the non-instance_type values should have
|
|
# been touched. Since we didn't set created_at on any
|
|
# of them, they should all still be None.
|
|
self.assertEqual(result['value'], original['value'])
|
|
self.assertEqual(result['created_at'], None)
|
|
|
|
def _pre_upgrade_172(self, engine):
|
|
instance_types = db_utils.get_table(engine, 'instance_types')
|
|
data = [
|
|
dict(id=21, name='uc_name0', memory_mb=128, vcpus=1,
|
|
root_gb=10, ephemeral_gb=0, flavorid="uc_flavor1", swap=0,
|
|
rxtx_factor=1.0, vcpu_weight=1, disabled=False,
|
|
is_public=True, deleted=0),
|
|
dict(id=22, name='uc_name1', memory_mb=128, vcpus=1,
|
|
root_gb=10, ephemeral_gb=0, flavorid="uc_flavor1", swap=0,
|
|
rxtx_factor=1.0, vcpu_weight=1, disabled=False,
|
|
is_public=True, deleted=0),
|
|
dict(id=23, name='uc_name2', memory_mb=128, vcpus=1,
|
|
root_gb=10, ephemeral_gb=0, flavorid="uc_flavor2", swap=0,
|
|
rxtx_factor=1.0, vcpu_weight=1, disabled=False,
|
|
is_public=True, deleted=0),
|
|
dict(id=24, name='uc_name2', memory_mb=128, vcpus=1,
|
|
root_gb=10, ephemeral_gb=0, flavorid="uc_flavor3", swap=0,
|
|
rxtx_factor=1.0, vcpu_weight=1, disabled=False,
|
|
is_public=True, deleted=0),
|
|
]
|
|
engine.execute(instance_types.insert(), data)
|
|
return data
|
|
|
|
def _check_172(self, engine, data):
|
|
instance_types = db_utils.get_table(engine, 'instance_types')
|
|
|
|
not_deleted = instance_types.c.deleted != instance_types.c.id
|
|
|
|
# There is only one instance_type with flavor `uc_flavor1`
|
|
uc_flavor1_rows = instance_types.select().\
|
|
where(instance_types.c.flavorid == 'uc_flavor1').\
|
|
where(not_deleted).\
|
|
execute().\
|
|
fetchall()
|
|
|
|
self.assertEqual(1, len(uc_flavor1_rows))
|
|
|
|
# There is only one instance_type with name `uc_name2`
|
|
uc_name2_rows = instance_types.select().\
|
|
where(instance_types.c.name == 'uc_name2').\
|
|
where(not_deleted).\
|
|
execute().\
|
|
fetchall()
|
|
self.assertEqual(1, len(uc_name2_rows))
|
|
|
|
# migration 173, add unique constraint to keypairs
|
|
def _pre_upgrade_173(self, engine):
|
|
created_at = [timeutils.utcnow() for x in range(0, 7)]
|
|
fake_keypairs = [dict(name='key1', user_id='1a',
|
|
created_at=created_at[0],
|
|
deleted=0),
|
|
dict(name='key1', user_id='1a',
|
|
created_at=created_at[1],
|
|
deleted=0),
|
|
dict(name='key1', user_id='1a',
|
|
created_at=created_at[2],
|
|
deleted=0)
|
|
]
|
|
keypairs = db_utils.get_table(engine, 'key_pairs')
|
|
engine.execute(keypairs.insert(), fake_keypairs)
|
|
return fake_keypairs
|
|
|
|
def _check_173(self, engine, data):
|
|
keypairs = db_utils.get_table(engine, 'key_pairs')
|
|
# Unique constraints are not listed in table.constraints for any db.
|
|
# So, simply add a duplicate keypair to check if unique constraint
|
|
# is applied to the key_pairs table or not.
|
|
insert = keypairs.insert()
|
|
duplicate_keypair = dict(name='key4', user_id='4a',
|
|
created_at=timeutils.utcnow(),
|
|
deleted=0)
|
|
insert.execute(duplicate_keypair)
|
|
# Insert again
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError, insert.execute,
|
|
duplicate_keypair)
|
|
|
|
# Get all unique records
|
|
rows = keypairs.select().\
|
|
where(keypairs.c.deleted != keypairs.c.id).\
|
|
execute().\
|
|
fetchall()
|
|
# Ensure the number of unique keypairs is correct
|
|
self.assertEqual(len(rows), 2)
|
|
|
|
def _pre_upgrade_174(self, engine):
|
|
instance_types = db_utils.get_table(engine, 'instance_types')
|
|
instance_type_projects = db_utils.get_table(engine,
|
|
'instance_type_projects')
|
|
|
|
instance_type_data = [
|
|
dict(id=31, name='itp_name0', memory_mb=128, vcpus=1,
|
|
root_gb=10, ephemeral_gb=0, flavorid="itp_flavor1", swap=0,
|
|
rxtx_factor=1.0, vcpu_weight=1, disabled=False,
|
|
is_public=True, deleted=0)
|
|
]
|
|
instance_type_projects_data = [
|
|
dict(project_id='pr1', instance_type_id=31, deleted=0),
|
|
dict(project_id='pr1', instance_type_id=31, deleted=0),
|
|
dict(project_id='pr2', instance_type_id=31, deleted=0)
|
|
]
|
|
|
|
engine.execute(instance_types.insert(), instance_type_data)
|
|
engine.execute(instance_type_projects.insert(),
|
|
instance_type_projects_data)
|
|
|
|
def _check_174(self, engine, data):
|
|
it_projects = db_utils.get_table(engine, 'instance_type_projects')
|
|
|
|
def get_(project_id, it_id, deleted):
|
|
deleted_value = 0 if not deleted else it_projects.c.id
|
|
return it_projects.select().\
|
|
where(it_projects.c.project_id == project_id).\
|
|
where(it_projects.c.instance_type_id == it_id).\
|
|
where(it_projects.c.deleted == deleted_value).\
|
|
execute().\
|
|
fetchall()
|
|
|
|
self.assertEqual(1, len(get_('pr1', '31', False)))
|
|
self.assertEqual(1, len(get_('pr1', '31', True)))
|
|
self.assertEqual(1, len(get_('pr2', '31', False)))
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
it_projects.insert().execute,
|
|
dict(instance_type=31, project_id='pr1', deleted=0))
|
|
|
|
# migration 175, Modify volume_usage-cache, Drop column instance_id, add
|
|
# columns instance_uuid, project_id and user_id
|
|
def _pre_upgrade_175(self, engine):
|
|
volume_usage_cache = db_utils.get_table(engine, 'volume_usage_cache')
|
|
fake_usage = {'volume_id': 'fake_volume_id',
|
|
'instance_id': 10,
|
|
'tot_last_refreshed': timeutils.utcnow(),
|
|
'tot_reads': 2,
|
|
'tot_read_bytes': 3,
|
|
'tot_writes': 4,
|
|
'tot_write_bytes': 5,
|
|
'curr_last_refreshed': timeutils.utcnow(),
|
|
'curr_reads': 6,
|
|
'curr_read_bytes': 7,
|
|
'curr_writes': 8,
|
|
'curr_write_bytes': 9}
|
|
volume_usage_cache.insert().execute(fake_usage)
|
|
|
|
def _check_175(self, engine, data):
|
|
volume_usage_cache = db_utils.get_table(engine, 'volume_usage_cache')
|
|
# Get the record
|
|
rows = volume_usage_cache.select().execute().fetchall()
|
|
self.assertEqual(len(rows), 1)
|
|
|
|
self.assertEqual(rows[0]['instance_uuid'], None)
|
|
self.assertEqual(rows[0]['project_id'], None)
|
|
self.assertEqual(rows[0]['user_id'], None)
|
|
self.assertFalse('instance_id' in rows[0])
|
|
|
|
def _post_downgrade_175(self, engine):
|
|
volume_usage_cache = db_utils.get_table(engine, 'volume_usage_cache')
|
|
# Get the record
|
|
rows = volume_usage_cache.select().execute().fetchall()
|
|
self.assertEqual(len(rows), 1)
|
|
|
|
self.assertFalse('instance_uuid' in rows[0])
|
|
self.assertFalse('project_id' in rows[0])
|
|
self.assertFalse('user_id' in rows[0])
|
|
self.assertEqual(rows[0]['instance_id'], None)
|
|
|
|
def _check_176(self, engine, data):
|
|
volume_usage_cache = db_utils.get_table(engine, 'volume_usage_cache')
|
|
# Get the record
|
|
rows = volume_usage_cache.select().execute().fetchall()
|
|
self.assertEqual(len(rows), 1)
|
|
|
|
self.assertEqual(rows[0]['availability_zone'], None)
|
|
|
|
def _post_downgrade_176(self, engine):
|
|
volume_usage_cache = db_utils.get_table(engine, 'volume_usage_cache')
|
|
# Get the record
|
|
rows = volume_usage_cache.select().execute().fetchall()
|
|
self.assertEqual(len(rows), 1)
|
|
|
|
self.assertFalse('availability_zone' in rows[0])
|
|
|
|
def _pre_upgrade_177(self, engine):
|
|
floating_ips = db_utils.get_table(engine, 'floating_ips')
|
|
data = [
|
|
{'address': '128.128.128.128', 'deleted': 0},
|
|
{'address': '128.128.128.128', 'deleted': 0},
|
|
{'address': '128.128.128.129', 'deleted': 0},
|
|
]
|
|
|
|
for item in data:
|
|
floating_ips.insert().values(item).execute()
|
|
return data
|
|
|
|
def _check_177(self, engine, data):
|
|
floating_ips = db_utils.get_table(engine, 'floating_ips')
|
|
|
|
def get_(address, deleted):
|
|
deleted_value = 0 if not deleted else floating_ips.c.id
|
|
return floating_ips.select().\
|
|
where(floating_ips.c.address == address).\
|
|
where(floating_ips.c.deleted == deleted_value).\
|
|
execute().\
|
|
fetchall()
|
|
|
|
self.assertEqual(1, len(get_('128.128.128.128', False)))
|
|
self.assertEqual(1, len(get_('128.128.128.128', True)))
|
|
self.assertEqual(1, len(get_('128.128.128.129', False)))
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
floating_ips.insert().execute,
|
|
dict(address='128.128.128.129', deleted=0))
|
|
|
|
# migration 179 - convert cells.deleted from boolean to int
|
|
def _pre_upgrade_179(self, engine):
|
|
cells_data = [
|
|
{'id': 4, 'deleted': True},
|
|
{'id': 5, 'deleted': False},
|
|
]
|
|
|
|
cells = db_utils.get_table(engine, 'cells')
|
|
engine.execute(cells.insert(), cells_data)
|
|
|
|
return dict(cells=cells_data)
|
|
|
|
def _check_179(self, engine, data):
|
|
cells = db_utils.get_table(engine, 'cells')
|
|
cell = cells.select(cells.c.id == 4).execute().first()
|
|
self.assertEqual(4, cell.deleted)
|
|
cell = cells.select(cells.c.id == 5).execute().first()
|
|
self.assertEqual(0, cell.deleted)
|
|
|
|
def _check_180(self, engine, data):
|
|
self.assertTrue(db_utils.check_shadow_table(engine,
|
|
"volume_usage_cache"))
|
|
|
|
def _check_181(self, engine, data):
|
|
self.assertTrue(db_utils.check_shadow_table(engine, 'cells'))
|
|
|
|
def _pre_upgrade_182(self, engine):
|
|
CIDR = '6666:1020:1000:2013:1000:6535:abcd:abcd'
|
|
|
|
security_group_rules = \
|
|
db_utils.get_table(engine, 'shadow_security_group_rules')
|
|
values = {
|
|
'id': 182,
|
|
'protocol': 'tcp',
|
|
'from_port': 6666,
|
|
'to_port': 9999,
|
|
'cidr': CIDR,
|
|
'deleted': 0
|
|
}
|
|
security_group_rules.insert().values(values).execute()
|
|
|
|
networks = db_utils.get_table(engine, 'shadow_networks')
|
|
values = {
|
|
'id': 182,
|
|
'vlan': 100500,
|
|
'cidr': CIDR,
|
|
'cidr_v6': CIDR,
|
|
'deleted': 0
|
|
}
|
|
networks.insert().values(values).execute()
|
|
|
|
provider_fw_rules = db_utils.get_table(engine,
|
|
'shadow_provider_fw_rules')
|
|
values = {
|
|
'id': 182,
|
|
'protocol': 'tcp',
|
|
'from_port': 6666,
|
|
'to_port': 9999,
|
|
'cidr': CIDR,
|
|
'deleted': 0
|
|
}
|
|
provider_fw_rules.insert().values(values).execute()
|
|
return {'cidr': CIDR}
|
|
|
|
def _check_182(self, engine, data):
|
|
self.assertTrue(db_utils.check_shadow_table(engine,
|
|
'security_group_rules'))
|
|
self.assertTrue(db_utils.check_shadow_table(engine,
|
|
'provider_fw_rules'))
|
|
self.assertTrue(db_utils.check_shadow_table(engine, 'networks'))
|
|
|
|
table_fields = {
|
|
'shadow_security_group_rules': ['cidr'],
|
|
'shadow_networks': ['cidr', 'cidr_v6'],
|
|
'shadow_provider_fw_rules': ['cidr']
|
|
}
|
|
|
|
for table_name, fields in table_fields.iteritems():
|
|
table = db_utils.get_table(engine, table_name)
|
|
rows = table.\
|
|
select().\
|
|
where(table.c.id == 182).\
|
|
execute().\
|
|
fetchall()
|
|
self.assertEqual(len(rows), 1)
|
|
for field in fields:
|
|
self.assertEqual(rows[0][field], data['cidr'])
|
|
|
|
for field in fields:
|
|
# we should be able to store mask in cidr fields also
|
|
table.\
|
|
update().\
|
|
values({field: data['cidr'] + '/128'}).\
|
|
execute()
|
|
|
|
def _check_183(self, engine, data):
|
|
table_name = 'security_group_default_rules'
|
|
self.assertTrue(db_utils.check_shadow_table(engine, table_name))
|
|
|
|
def _check_184(self, engine, data):
|
|
self.assertTrue(db_utils.check_shadow_table(engine, 'instances'))
|
|
self.assertTrue(db_utils.check_shadow_table(engine, 'networks'))
|
|
self.assertTrue(db_utils.check_shadow_table(engine, 'fixed_ips'))
|
|
self.assertTrue(db_utils.check_shadow_table(engine, 'floating_ips'))
|
|
self.assertTrue(db_utils.check_shadow_table(engine, 'console_pools'))
|
|
|
|
def _unique_constraint_check_migrate_185(self, engine, check=True):
|
|
"""Test check unique constraint behavior.
|
|
|
|
It should be the same before and after migration because we
|
|
changed their names only.
|
|
"""
|
|
|
|
data_list = [
|
|
("floating_ips", {'address': '10.12.14.16', 'deleted': 0}),
|
|
("instance_info_caches", {'instance_uuid': 'm161-uuid1'}),
|
|
('instance_type_projects', {'instance_type_id': 1,
|
|
'project_id': '116', 'deleted': 0}),
|
|
('instance_types', {'flavorid': "flavorid_12", 'deleted': 0,
|
|
'memory_mb': 64, 'vcpus': 10, 'swap': 100}),
|
|
('instance_types', {'name': "name_123", 'deleted': 0,
|
|
'memory_mb': 128, 'vcpus': 11, 'swap': 300}),
|
|
('key_pairs', {'user_id': 1, 'name': "name_qwer", 'deleted': 0}),
|
|
('networks', {'vlan': '123', 'deleted': 0}),
|
|
('task_log', {'task_name': 'task_123', 'host': 'localhost',
|
|
'period_beginning': datetime.datetime(2013, 2, 11),
|
|
'period_ending': datetime.datetime(2015, 1, 1),
|
|
'state': 'state_1', 'message': 'msg_1'}),
|
|
('virtual_interfaces', {'address': '192.168.0.0'})
|
|
]
|
|
|
|
for table_name, data in data_list:
|
|
table = db_utils.get_table(engine, table_name)
|
|
if not check:
|
|
table.insert().values(data).execute()
|
|
else:
|
|
# we replace values for some columns because they don't
|
|
# belong to unique constraint
|
|
if table_name == "instance_types":
|
|
for key in ("memory_mb", "vcpus", "swap"):
|
|
data[key] = data[key] * 2
|
|
if table_name == "task_log":
|
|
data["message"] = 'msg_2'
|
|
data["state"] = 'state_2'
|
|
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
table.insert().execute, data)
|
|
|
|
def _pre_upgrade_185(self, engine):
|
|
self._unique_constraint_check_migrate_185(engine, False)
|
|
|
|
def check_185(self, engine):
|
|
self._unique_constraint_check_migrate_185(engine)
|
|
|
|
def _post_downgrade_185(self, engine):
|
|
self._unique_constraint_check_migrate_185(engine)
|
|
|
|
def _pre_upgrade_186(self, engine):
|
|
self.mig186_fake_instances = [
|
|
dict(uuid='mig186_uuid-1', image_ref='fake_image_1',
|
|
root_device_name='/dev/vda'),
|
|
dict(uuid='mig186_uuid-2', image_ref='',
|
|
root_device_name='vda'),
|
|
dict(uuid='mig186_uuid-3', image_ref='fake_image_2',
|
|
root_device_name='/dev/vda'),
|
|
]
|
|
|
|
self.mig186_fake_bdms = [
|
|
# Instance 1 - image, volume and swap
|
|
dict(instance_uuid='mig186_uuid-1', device_name='/dev/vdc',
|
|
volume_id='fake_volume_1'),
|
|
dict(instance_uuid='mig186_uuid-1', device_name='/dev/vdb',
|
|
virtual_name='swap'),
|
|
# Instance 2 - no image. snapshot and volume
|
|
dict(instance_uuid='mig186_uuid-2', device_name='/dev/vda',
|
|
snapshot_id='fake_snap_1', volume_id='fake_volume_2'),
|
|
dict(instance_uuid='mig186_uuid-2', device_name='/dev/vdc',
|
|
volume_id='fake_volume_3'),
|
|
# Instance 3 - ephemerals and swap
|
|
dict(instance_uuid='mig186_uuid-3', device_name='/dev/vdc',
|
|
virtual_name='ephemeral0'),
|
|
dict(instance_uuid='mig186_uuid-3', device_name='/dev/vdd',
|
|
virtual_name='ephemeral1'),
|
|
dict(instance_uuid='mig186_uuid-3', device_name='/dev/vdb',
|
|
virtual_name='swap'),
|
|
]
|
|
|
|
instances = db_utils.get_table(engine, 'instances')
|
|
block_device = db_utils.get_table(engine, 'block_device_mapping')
|
|
engine.execute(instances.insert(), self.mig186_fake_instances)
|
|
for fake_bdm in self.mig186_fake_bdms:
|
|
engine.execute(block_device.insert(), fake_bdm)
|
|
|
|
return self.mig186_fake_instances, self.mig186_fake_bdms
|
|
|
|
def _check_186(self, engine, data):
|
|
block_device = db_utils.get_table(engine, 'block_device_mapping')
|
|
|
|
instance_qs = []
|
|
|
|
for instance in ('mig186_uuid-1', 'mig186_uuid-2', 'mig186_uuid-3'):
|
|
q = block_device.select().where(
|
|
block_device.c.instance_uuid == instance).order_by(
|
|
block_device.c.id.asc()
|
|
)
|
|
instance_qs.append(q)
|
|
|
|
bdm_1s, bdm_2s, bdm_3s = (
|
|
[bdm for bdm in q.execute()]
|
|
for q in instance_qs
|
|
)
|
|
|
|
self.assertEqual(len(bdm_1s), 3)
|
|
self.assertEqual(len(bdm_2s), 2)
|
|
self.assertEqual(len(bdm_3s), 4)
|
|
|
|
# Instance 1
|
|
self.assertEqual(bdm_1s[0].source_type, 'volume')
|
|
self.assertEqual(bdm_1s[0].destination_type, 'volume')
|
|
self.assertEqual(bdm_1s[0].volume_id, 'fake_volume_1')
|
|
self.assertEqual(bdm_1s[0].device_type, 'disk')
|
|
self.assertEqual(bdm_1s[0].boot_index, -1)
|
|
self.assertEqual(bdm_1s[0].device_name, '/dev/vdc')
|
|
|
|
self.assertEqual(bdm_1s[1].source_type, 'blank')
|
|
self.assertEqual(bdm_1s[1].guest_format, 'swap')
|
|
self.assertEqual(bdm_1s[1].destination_type, 'local')
|
|
self.assertEqual(bdm_1s[1].device_type, 'disk')
|
|
self.assertEqual(bdm_1s[1].boot_index, -1)
|
|
self.assertEqual(bdm_1s[1].device_name, '/dev/vdb')
|
|
|
|
self.assertEqual(bdm_1s[2].source_type, 'image')
|
|
self.assertEqual(bdm_1s[2].destination_type, 'local')
|
|
self.assertEqual(bdm_1s[2].device_type, 'disk')
|
|
self.assertEqual(bdm_1s[2].image_id, 'fake_image_1')
|
|
self.assertEqual(bdm_1s[2].boot_index, 0)
|
|
|
|
# Instance 2
|
|
self.assertEqual(bdm_2s[0].source_type, 'snapshot')
|
|
self.assertEqual(bdm_2s[0].destination_type, 'volume')
|
|
self.assertEqual(bdm_2s[0].snapshot_id, 'fake_snap_1')
|
|
self.assertEqual(bdm_2s[0].volume_id, 'fake_volume_2')
|
|
self.assertEqual(bdm_2s[0].device_type, 'disk')
|
|
self.assertEqual(bdm_2s[0].boot_index, 0)
|
|
self.assertEqual(bdm_2s[0].device_name, '/dev/vda')
|
|
|
|
self.assertEqual(bdm_2s[1].source_type, 'volume')
|
|
self.assertEqual(bdm_2s[1].destination_type, 'volume')
|
|
self.assertEqual(bdm_2s[1].volume_id, 'fake_volume_3')
|
|
self.assertEqual(bdm_2s[1].device_type, 'disk')
|
|
self.assertEqual(bdm_2s[1].boot_index, -1)
|
|
self.assertEqual(bdm_2s[1].device_name, '/dev/vdc')
|
|
|
|
# Instance 3
|
|
self.assertEqual(bdm_3s[0].source_type, 'blank')
|
|
self.assertEqual(bdm_3s[0].destination_type, 'local')
|
|
self.assertEqual(bdm_3s[0].device_type, 'disk')
|
|
self.assertEqual(bdm_3s[0].boot_index, -1)
|
|
self.assertEqual(bdm_3s[0].device_name, '/dev/vdc')
|
|
|
|
self.assertEqual(bdm_3s[1].source_type, 'blank')
|
|
self.assertEqual(bdm_3s[1].destination_type, 'local')
|
|
self.assertEqual(bdm_3s[1].device_type, 'disk')
|
|
self.assertEqual(bdm_3s[1].boot_index, -1)
|
|
self.assertEqual(bdm_3s[1].device_name, '/dev/vdd')
|
|
|
|
self.assertEqual(bdm_3s[2].source_type, 'blank')
|
|
self.assertEqual(bdm_3s[2].guest_format, 'swap')
|
|
self.assertEqual(bdm_3s[2].destination_type, 'local')
|
|
self.assertEqual(bdm_3s[2].device_type, 'disk')
|
|
self.assertEqual(bdm_3s[2].boot_index, -1)
|
|
self.assertEqual(bdm_3s[2].device_name, '/dev/vdb')
|
|
|
|
self.assertEqual(bdm_3s[3].source_type, 'image')
|
|
self.assertEqual(bdm_3s[3].destination_type, 'local')
|
|
self.assertEqual(bdm_3s[3].device_type, 'disk')
|
|
self.assertEqual(bdm_3s[3].image_id, 'fake_image_2')
|
|
self.assertEqual(bdm_3s[3].boot_index, 0)
|
|
|
|
def _post_downgrade_186(self, engine):
|
|
block_device = db_utils.get_table(engine, 'block_device_mapping')
|
|
|
|
q = block_device.select().where(
|
|
sqlalchemy.or_(
|
|
block_device.c.instance_uuid == 'mig186_uuid-1',
|
|
block_device.c.instance_uuid == 'mig186_uuid-2',
|
|
block_device.c.instance_uuid == 'mig186_uuid-3'))\
|
|
.order_by(block_device.c.device_name.asc())
|
|
|
|
expected_bdms = sorted(self.mig186_fake_bdms,
|
|
key=operator.itemgetter('device_name'))
|
|
got_bdms = [bdm for bdm in q.execute()]
|
|
|
|
self.assertEquals(len(expected_bdms), len(got_bdms))
|
|
for expected, got in zip(expected_bdms, got_bdms):
|
|
self.assertThat(expected, matchers.IsSubDictOf(dict(got)))
|
|
|
|
# addition of the vm instance groups
|
|
def _check_no_group_instance_tables(self, engine):
|
|
self.assertRaises(sqlalchemy.exc.NoSuchTableError,
|
|
db_utils.get_table, engine,
|
|
'instance_groups')
|
|
self.assertRaises(sqlalchemy.exc.NoSuchTableError,
|
|
db_utils.get_table, engine,
|
|
'instance_group_member')
|
|
self.assertRaises(sqlalchemy.exc.NoSuchTableError,
|
|
db_utils.get_table, engine,
|
|
'instance_group_policy')
|
|
self.assertRaises(sqlalchemy.exc.NoSuchTableError,
|
|
db_utils.get_table, engine,
|
|
'instance_group_metadata')
|
|
|
|
def _check_group_instance_groups(self, engine):
|
|
groups = db_utils.get_table(engine, 'instance_groups')
|
|
uuid4 = uuidutils.generate_uuid()
|
|
uuid5 = uuidutils.generate_uuid()
|
|
group_data = [
|
|
{'id': 4, 'deleted': 4, 'uuid': uuid4},
|
|
{'id': 5, 'deleted': 0, 'uuid': uuid5},
|
|
]
|
|
engine.execute(groups.insert(), group_data)
|
|
group = groups.select(groups.c.id == 4).execute().first()
|
|
self.assertEqual(4, group.deleted)
|
|
group = groups.select(groups.c.id == 5).execute().first()
|
|
self.assertEqual(0, group.deleted)
|
|
|
|
def _pre_upgrade_187(self, engine):
|
|
self._check_no_group_instance_tables(engine)
|
|
|
|
def _check_187(self, engine, data):
|
|
self._check_group_instance_groups(engine)
|
|
tables = ['instance_group_policy', 'instance_group_metadata',
|
|
'instance_group_member']
|
|
for table in tables:
|
|
db_utils.get_table(engine, table)
|
|
|
|
def _post_downgrade_187(self, engine):
|
|
# check that groups does not exist
|
|
self._check_no_group_instance_tables(engine)
|
|
|
|
def _check_188(self, engine, data):
|
|
services = db_utils.get_table(engine, 'services')
|
|
rows = services.select().execute().fetchall()
|
|
self.assertEqual(rows[0]['disabled_reason'], None)
|
|
|
|
def _post_downgrade_188(self, engine):
|
|
services = db_utils.get_table(engine, 'services')
|
|
rows = services.select().execute().fetchall()
|
|
self.assertFalse('disabled_reason' in rows[0])
|
|
|
|
def _pre_upgrade_189(self, engine):
|
|
cells = db_utils.get_table(engine, 'cells')
|
|
data = [
|
|
{'name': 'name_123', 'deleted': 0},
|
|
{'name': 'name_123', 'deleted': 0},
|
|
{'name': 'name_345', 'deleted': 0},
|
|
]
|
|
for item in data:
|
|
cells.insert().values(item).execute()
|
|
return data
|
|
|
|
def _check_189(self, engine, data):
|
|
cells = db_utils.get_table(engine, 'cells')
|
|
|
|
def get_(name, deleted):
|
|
deleted_value = 0 if not deleted else cells.c.id
|
|
return cells.select().\
|
|
where(cells.c.name == name).\
|
|
where(cells.c.deleted == deleted_value).\
|
|
execute().\
|
|
fetchall()
|
|
|
|
self.assertEqual(1, len(get_('name_123', False)))
|
|
self.assertEqual(1, len(get_('name_123', True)))
|
|
self.assertEqual(1, len(get_('name_345', False)))
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
cells.insert().execute,
|
|
{'name': 'name_123', 'deleted': 0})
|
|
|
|
def _pre_upgrade_190(self, engine):
|
|
security_groups = db_utils.get_table(engine, 'security_groups')
|
|
data = [
|
|
{'name': 'group1', 'project_id': 'fake', 'deleted': 0},
|
|
{'name': 'group1', 'project_id': 'fake', 'deleted': 0},
|
|
{'name': 'group2', 'project_id': 'fake', 'deleted': 0},
|
|
]
|
|
|
|
for item in data:
|
|
security_groups.insert().values(item).execute()
|
|
|
|
def _check_190(self, engine, data):
|
|
security_groups = db_utils.get_table(engine, 'security_groups')
|
|
|
|
def get_(name, project_id, deleted):
|
|
deleted_value = 0 if not deleted else security_groups.c.id
|
|
return security_groups.select().\
|
|
where(security_groups.c.name == name).\
|
|
where(security_groups.c.project_id == project_id).\
|
|
where(security_groups.c.deleted == deleted_value).\
|
|
execute().\
|
|
fetchall()
|
|
|
|
self.assertEqual(1, len(get_('group1', 'fake', False)))
|
|
self.assertEqual(1, len(get_('group1', 'fake', True)))
|
|
self.assertEqual(1, len(get_('group2', 'fake', False)))
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
security_groups.insert().execute,
|
|
dict(name='group2', project_id='fake', deleted=0))
|
|
|
|
def _pre_upgrade_191(self, engine):
|
|
quotas = db_utils.get_table(engine, 'quotas')
|
|
data = [
|
|
{'project_id': 'project1', 'resource': 'resource1', 'deleted': 0},
|
|
{'project_id': 'project1', 'resource': 'resource1', 'deleted': 0},
|
|
{'project_id': 'project2', 'resource': 'resource1', 'deleted': 0},
|
|
]
|
|
for item in data:
|
|
quotas.insert().values(item).execute()
|
|
return data
|
|
|
|
def _check_191(self, engine, data):
|
|
quotas = db_utils.get_table(engine, 'quotas')
|
|
|
|
def get_(project_id, deleted):
|
|
deleted_value = 0 if not deleted else quotas.c.id
|
|
return quotas.select().\
|
|
where(quotas.c.project_id == project_id).\
|
|
where(quotas.c.deleted == deleted_value).\
|
|
execute().\
|
|
fetchall()
|
|
|
|
self.assertEqual(1, len(get_('project1', False)))
|
|
self.assertEqual(1, len(get_('project1', True)))
|
|
self.assertEqual(1, len(get_('project2', False)))
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
quotas.insert().execute,
|
|
{'project_id': 'project1', 'resource': 'resource1',
|
|
'deleted': 0})
|
|
|
|
def _check_192(self, engine, data):
|
|
virtual_if = db_utils.get_table(engine, 'virtual_interfaces')
|
|
values = {'address': 'address0', 'deleted': 0}
|
|
virtual_if.insert().values(values).execute()
|
|
values['deleted'] = 1
|
|
virtual_if.insert().values(values).execute()
|
|
values['deleted'] = 0
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
virtual_if.insert().execute,
|
|
values)
|
|
|
|
def _post_downgrade_192(self, engine):
|
|
virtual_if = db_utils.get_table(engine, 'virtual_interfaces')
|
|
deleted = virtual_if.select().\
|
|
where(virtual_if.c.deleted != 0).\
|
|
execute().fetchall()
|
|
self.assertEqual([], deleted)
|
|
values = {'address': 'address1'}
|
|
virtual_if.insert().values(values).execute()
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
virtual_if.insert().execute,
|
|
values)
|
|
|
|
def _check_193(self, engine, data):
|
|
tables = set(engine.table_names())
|
|
dropped_tables = set([
|
|
'sm_volume',
|
|
'sm_backend_config',
|
|
'sm_flavors',
|
|
'virtual_storage_arrays',
|
|
'volume_metadata',
|
|
'volume_type_extra_specs',
|
|
'volume_types',
|
|
'shadow_sm_volume',
|
|
'shadow_sm_backend_config',
|
|
'shadow_sm_flavors',
|
|
'shadow_virtual_storage_arrays',
|
|
'shadow_volume_metadata',
|
|
'shadow_volume_type_extra_specs',
|
|
'shadow_volume_types'])
|
|
check = bool(tables & dropped_tables)
|
|
self.assertFalse(check)
|
|
|
|
def _post_downgrade_193(self, engine):
|
|
tables = set(engine.table_names())
|
|
dropped_tables = set([
|
|
'sm_volume',
|
|
'sm_backend_config',
|
|
'sm_flavors',
|
|
'virtual_storage_arrays',
|
|
'volume_metadata',
|
|
'volume_type_extra_specs',
|
|
'volume_types',
|
|
'shadow_sm_volume',
|
|
'shadow_sm_backend_config',
|
|
'shadow_sm_flavors',
|
|
'shadow_virtual_storage_arrays',
|
|
'shadow_volume_metadata',
|
|
'shadow_volume_type_extra_specs',
|
|
'shadow_volume_types'])
|
|
check = tables & dropped_tables
|
|
self.assertEqual(check, dropped_tables)
|
|
|
|
def _check_194(self, engine, data):
|
|
if engine.name == 'sqlite':
|
|
return
|
|
|
|
test_data = {
|
|
# table_name: ((index_name_1, (*columns)), ...)
|
|
"certificates": (
|
|
("certificates_project_id_deleted_idx",
|
|
["project_id", "deleted"]),
|
|
("certificates_user_id_deleted_idx", ["user_id", "deleted"]),
|
|
),
|
|
"instances": (
|
|
("instances_host_deleted_idx", ["host", "deleted"]),
|
|
("instances_uuid_deleted_idx", ["uuid", "deleted"]),
|
|
("instances_host_node_deleted_idx",
|
|
["host", "node", "deleted"]),
|
|
),
|
|
"iscsi_targets": (
|
|
("iscsi_targets_host_volume_id_deleted_idx",
|
|
["host", "volume_id", "deleted"]),
|
|
),
|
|
"networks": (
|
|
("networks_bridge_deleted_idx", ["bridge", "deleted"]),
|
|
("networks_project_id_deleted_idx", ["project_id", "deleted"]),
|
|
("networks_uuid_project_id_deleted_idx",
|
|
["uuid", "project_id", "deleted"]),
|
|
("networks_vlan_deleted_idx", ["vlan", "deleted"]),
|
|
),
|
|
"fixed_ips": (
|
|
("fixed_ips_network_id_host_deleted_idx",
|
|
["network_id", "host", "deleted"]),
|
|
("fixed_ips_address_reserved_network_id_deleted_idx",
|
|
["address", "reserved", "network_id", "deleted"]),
|
|
("fixed_ips_deleted_allocated_idx",
|
|
["address", "deleted", "allocated"]),
|
|
),
|
|
"floating_ips": (
|
|
("floating_ips_pool_deleted_fixed_ip_id_project_id_idx",
|
|
["pool", "deleted", "fixed_ip_id", "project_id"]),
|
|
),
|
|
"instance_faults": (
|
|
("instance_faults_instance_uuid_deleted_created_at_idx",
|
|
["instance_uuid", "deleted", "created_at"]),
|
|
),
|
|
}
|
|
|
|
for table_name, indexes in test_data.iteritems():
|
|
meta = sqlalchemy.MetaData()
|
|
meta.bind = engine
|
|
table = sqlalchemy.Table(table_name, meta, autoload=True)
|
|
|
|
index_data = [(idx.name, idx.columns.keys())
|
|
for idx in table.indexes]
|
|
|
|
for name, columns in indexes:
|
|
if engine.name == "postgresql":
|
|
# we can not get correct order of columns in index
|
|
# definition to postgresql using sqlalchemy. So we sortind
|
|
# columns list before compare
|
|
# bug http://www.sqlalchemy.org/trac/ticket/2767
|
|
self.assertIn(
|
|
(name, sorted(columns)),
|
|
([(idx[0], sorted(idx[1])) for idx in index_data])
|
|
)
|
|
else:
|
|
self.assertIn((name, columns), index_data)
|
|
|
|
def _pre_upgrade_195(self, engine):
|
|
fixed_ips = db_utils.get_table(engine, 'fixed_ips')
|
|
data = [
|
|
{'address': '128.128.128.128', 'deleted': 0},
|
|
{'address': '128.128.128.128', 'deleted': 0},
|
|
{'address': '128.128.128.129', 'deleted': 0},
|
|
]
|
|
|
|
for item in data:
|
|
fixed_ips.insert().values(item).execute()
|
|
return data
|
|
|
|
def _check_195(self, engine, data):
|
|
fixed_ips = db_utils.get_table(engine, 'fixed_ips')
|
|
|
|
def get_(address, deleted):
|
|
deleted_value = 0 if not deleted else fixed_ips.c.id
|
|
return fixed_ips.select()\
|
|
.where(fixed_ips.c.address == address)\
|
|
.where(fixed_ips.c.deleted == deleted_value)\
|
|
.execute()\
|
|
.fetchall()
|
|
|
|
self.assertEqual(1, len(get_('128.128.128.128', False)))
|
|
self.assertEqual(1, len(get_('128.128.128.128', True)))
|
|
self.assertEqual(1, len(get_('128.128.128.129', False)))
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
fixed_ips.insert().execute,
|
|
dict(address='128.128.128.129', deleted=0))
|
|
|
|
def _pre_upgrade_196(self, engine):
|
|
services = db_utils.get_table(engine, 'services')
|
|
data = [
|
|
{'id': 6, 'host': 'host1', 'binary': 'binary1', 'topic': 'topic3',
|
|
'report_count': 5, 'deleted': 0},
|
|
{'id': 7, 'host': 'host1', 'binary': 'binary1', 'topic': 'topic4',
|
|
'report_count': 5, 'deleted': 0},
|
|
{'id': 8, 'host': 'host2', 'binary': 'binary7', 'topic': 'topic2',
|
|
'report_count': 5, 'deleted': 0},
|
|
{'id': 9, 'host': 'host2', 'binary': 'binary8', 'topic': 'topic2',
|
|
'report_count': 5, 'deleted': 0},
|
|
]
|
|
for item in data:
|
|
services.insert().values(item).execute()
|
|
return data
|
|
|
|
def _check_196(self, engine, data):
|
|
services = db_utils.get_table(engine, 'services')
|
|
|
|
def get_(host, deleted):
|
|
deleted_value = 0 if not deleted else services.c.id
|
|
return services.select().\
|
|
where(services.c.host == host).\
|
|
where(services.c.deleted == deleted_value).\
|
|
execute().\
|
|
fetchall()
|
|
|
|
self.assertEqual(1, len(get_('host1', False)))
|
|
self.assertEqual(1, len(get_('host1', True)))
|
|
self.assertEqual(1, len(get_('host2', False)))
|
|
self.assertEqual(1, len(get_('host2', True)))
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
services.insert().execute,
|
|
{'id': 10, 'host': 'host1', 'binary': 'binary1',
|
|
'report_count': 5, 'topic': 'topic0', 'deleted': 0})
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
services.insert().execute,
|
|
{'id': 11, 'host': 'host2', 'binary': 'binary0',
|
|
'report_count': 5, 'topic': 'topic2', 'deleted': 0})
|
|
|
|
def _pre_upgrade_197(self, engine):
|
|
abuilds = db_utils.get_table(engine, 'agent_builds')
|
|
data = [
|
|
{'hypervisor': 'hyper1', 'os': 'os1', 'architecture': 'arch1',
|
|
'deleted': 0},
|
|
{'hypervisor': 'hyper1', 'os': 'os1', 'architecture': 'arch1',
|
|
'deleted': 0},
|
|
{'hypervisor': 'hyper2', 'os': 'os1', 'architecture': 'arch1',
|
|
'deleted': 0},
|
|
]
|
|
for item in data:
|
|
abuilds.insert().values(item).execute()
|
|
return data
|
|
|
|
def _check_197(self, engine, data):
|
|
abuilds = db_utils.get_table(engine, 'agent_builds')
|
|
|
|
def get_(hypervisor, deleted):
|
|
deleted_value = 0 if not deleted else abuilds.c.id
|
|
return abuilds.select().\
|
|
where(abuilds.c.hypervisor == hypervisor).\
|
|
where(abuilds.c.deleted == deleted_value).\
|
|
execute().\
|
|
fetchall()
|
|
|
|
self.assertEqual(1, len(get_('hyper1', False)))
|
|
self.assertEqual(1, len(get_('hyper1', True)))
|
|
self.assertEqual(1, len(get_('hyper2', False)))
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
abuilds.insert().execute,
|
|
{'hypervisor': 'hyper2', 'os': 'os1',
|
|
'architecture': 'arch1', 'deleted': 0})
|
|
|
|
def _pre_upgrade_198(self, engine):
|
|
table = db_utils.get_table(engine, 'console_pools')
|
|
data = [
|
|
{'host': 'localhost', 'console_type': 'type_1',
|
|
'compute_host': '192.168.122.8', 'deleted': 0},
|
|
{'host': 'localhost', 'console_type': 'type_1',
|
|
'compute_host': '192.168.122.8', 'deleted': 0},
|
|
{'host': 'localhost', 'console_type': 'type_2',
|
|
'compute_host': '192.168.122.8', 'deleted': 0},
|
|
]
|
|
|
|
for item in data:
|
|
table.insert().values(item).execute()
|
|
return data
|
|
|
|
def _check_198(self, engine, data):
|
|
table = db_utils.get_table(engine, 'console_pools')
|
|
|
|
def get_(host, console_type, compute_host, deleted):
|
|
deleted_value = 0 if not deleted else table.c.id
|
|
return table.select()\
|
|
.where(table.c.host == host)\
|
|
.where(table.c.console_type == console_type)\
|
|
.where(table.c.compute_host == compute_host)\
|
|
.where(table.c.deleted == deleted_value)\
|
|
.execute()\
|
|
.fetchall()
|
|
|
|
self.assertEqual(
|
|
1, len(get_('localhost', 'type_1', '192.168.122.8', False)))
|
|
self.assertEqual(
|
|
1, len(get_('localhost', 'type_1', '192.168.122.8', True)))
|
|
self.assertEqual(
|
|
1, len(get_('localhost', 'type_2', '192.168.122.8', False)))
|
|
self.assertRaises(
|
|
sqlalchemy.exc.IntegrityError,
|
|
table.insert().execute,
|
|
{'host': 'localhost', 'console_type': 'type_2',
|
|
'compute_host': '192.168.122.8', 'deleted': 0},
|
|
)
|
|
|
|
def _pre_upgrade_199(self, engine):
|
|
fake_aggregates = [{'id': 5, 'name': 'name1'},
|
|
{'id': 6, 'name': 'name2'}]
|
|
aggregates = db_utils.get_table(engine, 'aggregates')
|
|
engine.execute(aggregates.insert(), fake_aggregates)
|
|
|
|
fake_metadata = [{'aggregate_id': 5, 'key': 'availability_zone',
|
|
'value': 'custom_az'},
|
|
{'aggregate_id': 6, 'key': 'availability_zone',
|
|
'value': 'custom_az'}]
|
|
metadata = db_utils.get_table(engine, 'aggregate_metadata')
|
|
engine.execute(metadata.insert(), fake_metadata)
|
|
|
|
ahosts = db_utils.get_table(engine, 'aggregate_hosts')
|
|
data = [
|
|
{'host': 'host1', 'aggregate_id': 5, 'deleted': 0},
|
|
{'host': 'host1', 'aggregate_id': 5, 'deleted': 0},
|
|
{'host': 'host2', 'aggregate_id': 6, 'deleted': 0},
|
|
]
|
|
for item in data:
|
|
ahosts.insert().values(item).execute()
|
|
return data
|
|
|
|
def _check_199(self, engine, data):
|
|
ahosts = db_utils.get_table(engine, 'aggregate_hosts')
|
|
|
|
def get_(host, deleted):
|
|
deleted_value = 0 if not deleted else ahosts.c.id
|
|
return ahosts.select().\
|
|
where(ahosts.c.host == host).\
|
|
where(ahosts.c.deleted == deleted_value).\
|
|
execute().\
|
|
fetchall()
|
|
|
|
self.assertEqual(1, len(get_('host1', False)))
|
|
self.assertEqual(1, len(get_('host1', True)))
|
|
self.assertEqual(1, len(get_('host2', False)))
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
ahosts.insert().execute,
|
|
{'host': 'host2', 'aggregate_id': 6,
|
|
'deleted': 0})
|
|
|
|
def _pre_upgrade_200(self, engine):
|
|
cells = db_utils.get_table(engine, 'cells')
|
|
shadow_cells = db_utils.get_table(engine, 'shadow_cells')
|
|
cells.delete().execute()
|
|
shadow_cells.delete().execute()
|
|
data = [
|
|
{
|
|
'name': 'cell_transport_123',
|
|
'deleted': 0,
|
|
'username': 'user_123',
|
|
'password': 'pass_123',
|
|
'rpc_host': 'host_123',
|
|
'rpc_port': 123,
|
|
'rpc_virtual_host': 'virtual_123',
|
|
},
|
|
{
|
|
'name': 'cell_transport_456',
|
|
'deleted': 0,
|
|
'username': 'user_456',
|
|
'password': 'pass_456',
|
|
'rpc_host': 'ffff::456',
|
|
'rpc_port': 456,
|
|
'rpc_virtual_host': 'virtual_456',
|
|
},
|
|
]
|
|
for item in data:
|
|
cells.insert().values(item).execute()
|
|
shadow_cells.insert().values(item).execute()
|
|
return data
|
|
|
|
def _check_200(self, engine, data):
|
|
cells = db_utils.get_table(engine, 'cells')
|
|
shadow_cells = db_utils.get_table(engine, 'shadow_cells')
|
|
|
|
def get_(table, name):
|
|
return table.select().\
|
|
where(cells.c.name == name).\
|
|
execute().\
|
|
first()
|
|
|
|
for table in (cells, shadow_cells):
|
|
cell_123 = get_(cells, 'cell_transport_123')
|
|
self.assertNotIn('username', cell_123)
|
|
self.assertNotIn('password', cell_123)
|
|
self.assertNotIn('rpc_host', cell_123)
|
|
self.assertNotIn('rpc_port', cell_123)
|
|
self.assertNotIn('rpc_virtual_host', cell_123)
|
|
self.assertEqual('rabbit://user_123:pass_123@host_123:123/'
|
|
'virtual_123', cell_123['transport_url'])
|
|
|
|
cell_456 = get_(cells, 'cell_transport_456')
|
|
self.assertNotIn('username', cell_456)
|
|
self.assertNotIn('password', cell_456)
|
|
self.assertNotIn('rpc_host', cell_456)
|
|
self.assertNotIn('rpc_port', cell_456)
|
|
self.assertNotIn('rpc_virtual_host', cell_456)
|
|
self.assertEqual('rabbit://user_456:pass_456@[ffff::456]:456/'
|
|
'virtual_456', cell_456['transport_url'])
|
|
|
|
# Verify that the unique constraint still exists
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
cells.insert().execute,
|
|
{'name': 'cell_transport_123', 'deleted': 0})
|
|
|
|
def _post_downgrade_200(self, engine):
|
|
cells = db_utils.get_table(engine, 'cells')
|
|
shadow_cells = db_utils.get_table(engine, 'shadow_cells')
|
|
|
|
def get_(table, name):
|
|
return table.select().\
|
|
where(cells.c.name == name).\
|
|
execute().\
|
|
first()
|
|
|
|
for table in (cells, shadow_cells):
|
|
cell_123 = get_(cells, 'cell_transport_123')
|
|
self.assertNotIn('transport_url', cell_123)
|
|
self.assertEqual('user_123', cell_123['username'])
|
|
self.assertEqual('pass_123', cell_123['password'])
|
|
self.assertEqual('host_123', cell_123['rpc_host'])
|
|
self.assertEqual(123, cell_123['rpc_port'])
|
|
self.assertEqual('virtual_123', cell_123['rpc_virtual_host'])
|
|
|
|
cell_456 = get_(cells, 'cell_transport_456')
|
|
self.assertNotIn('transport_url', cell_456)
|
|
self.assertEqual('user_456', cell_456['username'])
|
|
self.assertEqual('pass_456', cell_456['password'])
|
|
self.assertEqual('ffff::456', cell_456['rpc_host'])
|
|
self.assertEqual(456, cell_456['rpc_port'])
|
|
self.assertEqual('virtual_456', cell_456['rpc_virtual_host'])
|
|
|
|
# Verify that the unique constraint still exists
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
cells.insert().execute,
|
|
{'name': 'cell_transport_123', 'deleted': 0})
|
|
|
|
def _check_201(self, engine, data):
|
|
if engine.name != 'sqlite':
|
|
return
|
|
|
|
data = {
|
|
# table_name: ((idx_1, (c1, c2,)), (idx2, (c1, c2,)), ...)
|
|
'agent_builds': (
|
|
('agent_builds_hypervisor_os_arch_idx',
|
|
('hypervisor', 'os', 'architecture'),),
|
|
),
|
|
'aggregate_metadata': (
|
|
('aggregate_metadata_key_idx', ('key',),),
|
|
),
|
|
'block_device_mapping': (
|
|
('block_device_mapping_instance_uuid_idx',
|
|
('instance_uuid',),),
|
|
('block_device_mapping_instance_uuid_device_name_idx',
|
|
('instance_uuid', 'device_name',),),
|
|
('block_device_mapping_instance_uuid_volume_id_idx',
|
|
('instance_uuid', 'volume_id',)),
|
|
('snapshot_id', ('snapshot_id',)),
|
|
('volume_id', ('volume_id',)),
|
|
),
|
|
'bw_usage_cache': (
|
|
('bw_usage_cache_uuid_start_period_idx',
|
|
('uuid', 'start_period',)),
|
|
),
|
|
'certificates': (
|
|
('certificates_project_id_deleted_idx',
|
|
('project_id', 'deleted',)),
|
|
('certificates_user_id_deleted_idx', ('user_id', 'deleted',)),
|
|
),
|
|
'compute_node_stats': (
|
|
('ix_compute_node_stats_compute_node_id',
|
|
('compute_node_id',)),
|
|
),
|
|
'consoles': (
|
|
('consoles_instance_uuid_idx', ('instance_uuid',)),
|
|
),
|
|
'dns_domains': (
|
|
('dns_domains_domain_deleted_idx', ('domain', 'deleted',)),
|
|
('project_id', ('project_id',)),
|
|
),
|
|
'fixed_ips': (
|
|
('address', ('address',)),
|
|
('fixed_ips_host_idx', ('host',)),
|
|
('fixed_ips_network_id_host_deleted_idx',
|
|
('network_id', 'host', 'deleted',)),
|
|
('fixed_ips_address_reserved_network_id_deleted_idx',
|
|
('address', 'reserved', 'network_id', 'deleted',)),
|
|
('network_id', ('network_id',)),
|
|
('fixed_ips_virtual_interface_id_fkey',
|
|
('virtual_interface_id',)),
|
|
('fixed_ips_instance_uuid_fkey', ('instance_uuid',)),
|
|
),
|
|
'floating_ips': (
|
|
('fixed_ip_id', ('fixed_ip_id',)),
|
|
('floating_ips_host_idx', ('host',)),
|
|
('floating_ips_project_id_idx', ('project_id',)),
|
|
('floating_ips_pool_deleted_fixed_ip_id_project_id_idx',
|
|
('pool', 'deleted', 'fixed_ip_id', 'project_id',)),
|
|
),
|
|
'instance_group_member': (
|
|
('instance_group_member_instance_idx', ('instance_id',)),
|
|
),
|
|
'instance_group_metadata': (
|
|
('instance_instance_group_metadata_key_idx', ('key',)),
|
|
),
|
|
'instance_group_policy': (
|
|
('instance_instance_group_policy_policy_idx', ('policy',)),
|
|
),
|
|
'instance_faults': (
|
|
('instance_faults_instance_uuid_deleted_created_at_idx',
|
|
('instance_uuid', 'deleted', 'created_at',)),
|
|
),
|
|
'instance_id_mappings': (
|
|
('ix_instance_id_mappings_uuid', ('uuid',)),
|
|
),
|
|
'instance_type_extra_specs': (
|
|
('instance_type_extra_specs_instance_type_id_key_idx',
|
|
('instance_type_id', 'key',)),
|
|
),
|
|
'instance_system_metadata': (
|
|
('instance_uuid', ('instance_uuid',)),
|
|
),
|
|
'instance_metadata': (
|
|
('instance_metadata_instance_uuid_idx', ('instance_uuid',)),
|
|
),
|
|
'instance_type_projects': (
|
|
('instance_type_id', ('instance_type_id',)),
|
|
),
|
|
'instances': (
|
|
('uuid', ('uuid',)),
|
|
),
|
|
'iscsi_targets': (
|
|
('iscsi_targets_host_idx', ('host',)),
|
|
('iscsi_targets_volume_id_fkey', ('volume_id',)),
|
|
('iscsi_targets_host_volume_id_deleted_idx',
|
|
('host', 'volume_id', 'deleted',)),
|
|
),
|
|
'networks': (
|
|
('networks_bridge_deleted_idx', ('bridge', 'deleted',)),
|
|
('networks_host_idx', ('host',)),
|
|
('networks_project_id_deleted_idx', ('project_id',
|
|
'deleted',)),
|
|
('networks_uuid_project_id_deleted_idx', ('uuid', 'project_id',
|
|
'deleted',)),
|
|
('networks_vlan_deleted_idx', ('vlan', 'deleted',)),
|
|
('networks_cidr_v6_idx', ('cidr_v6',)),
|
|
),
|
|
'reservations': (
|
|
('ix_reservations_project_id', ('project_id',)),
|
|
('usage_id', ('usage_id',)),
|
|
),
|
|
'security_group_instance_association': (
|
|
('security_group_instance_association_instance_uuid_idx',
|
|
('instance_uuid',)),
|
|
),
|
|
'quota_classes': (
|
|
('ix_quota_classes_class_name', ('class_name',)),
|
|
),
|
|
'quota_usages': (
|
|
('ix_quota_usages_project_id', ('project_id',)),
|
|
),
|
|
'virtual_interfaces': (
|
|
('virtual_interfaces_network_id_idx', ('network_id',)),
|
|
('virtual_interfaces_instance_uuid_fkey', ('instance_uuid',)),
|
|
),
|
|
'volumes': (
|
|
('volumes_instance_uuid_idx', ('instance_uuid',)),
|
|
),
|
|
'task_log': (
|
|
('ix_task_log_period_beginning', ('period_beginning',)),
|
|
('ix_task_log_host', ('host',)),
|
|
('ix_task_log_period_ending', ('period_ending',)),
|
|
),
|
|
}
|
|
|
|
meta = sqlalchemy.MetaData()
|
|
meta.bind = engine
|
|
|
|
for table_name, indexes in data.iteritems():
|
|
table = sqlalchemy.Table(table_name, meta, autoload=True)
|
|
indexes = [(i.name, tuple(i.columns.keys()))
|
|
for i in table.indexes]
|
|
|
|
for index in indexes:
|
|
self.assertIn(index, indexes)
|
|
|
|
def _pre_upgrade_202(self, engine):
|
|
fake_types = [
|
|
{'id': 35, 'name': 'type1', 'memory_mb': 128, 'vcpus': 1,
|
|
'root_gb': 10, 'ephemeral_gb': 0, 'flavorid': "1", 'swap': 0,
|
|
'rxtx_factor': 1.0, 'vcpu_weight': 1, 'disabled': False,
|
|
'is_public': True},
|
|
{'id': 36, 'name': 'type2', 'memory_mb': 512, 'vcpus': 1,
|
|
'root_gb': 10, 'ephemeral_gb': 5, 'flavorid': "2", 'swap': 0,
|
|
'rxtx_factor': 1.5, 'vcpu_weight': 2, 'disabled': False,
|
|
'is_public': True},
|
|
]
|
|
instance_types = db_utils.get_table(engine, 'instance_types')
|
|
engine.execute(instance_types.insert(), fake_types)
|
|
specs = db_utils.get_table(engine, 'instance_type_extra_specs')
|
|
data = [
|
|
{'instance_type_id': 35, 'key': 'key1', 'deleted': 0},
|
|
{'instance_type_id': 35, 'key': 'key1', 'deleted': 0},
|
|
{'instance_type_id': 36, 'key': 'key1', 'deleted': 0},
|
|
]
|
|
for item in data:
|
|
specs.insert().values(item).execute()
|
|
return data
|
|
|
|
def _check_202(self, engine, data):
|
|
specs = db_utils.get_table(engine, 'instance_type_extra_specs')
|
|
|
|
def get_(typeid, deleted):
|
|
deleted_value = 0 if not deleted else specs.c.id
|
|
return specs.select().\
|
|
where(specs.c.instance_type_id == typeid).\
|
|
where(specs.c.deleted == deleted_value).\
|
|
execute().\
|
|
fetchall()
|
|
|
|
self.assertEqual(1, len(get_(35, False)))
|
|
self.assertEqual(1, len(get_(35, True)))
|
|
self.assertEqual(1, len(get_(36, False)))
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
specs.insert().execute,
|
|
{'instance_type_id': 35, 'key': 'key1',
|
|
'deleted': 0})
|
|
|
|
# migration 203 - make user quotas key and value
|
|
def _pre_upgrade_203(self, engine):
|
|
quota_usages = db_utils.get_table(engine, 'quota_usages')
|
|
reservations = db_utils.get_table(engine, 'reservations')
|
|
fake_quota_usages = {'id': 5,
|
|
'resource': 'instances',
|
|
'in_use': 1,
|
|
'reserved': 1}
|
|
fake_reservations = {'id': 6,
|
|
'uuid': 'fake_reservationo_uuid',
|
|
'usage_id': 5,
|
|
'resource': 'instances',
|
|
'delta': 1,
|
|
'expire': timeutils.utcnow()}
|
|
quota_usages.insert().execute(fake_quota_usages)
|
|
reservations.insert().execute(fake_reservations)
|
|
|
|
def _check_203(self, engine, data):
|
|
project_user_quotas = db_utils.get_table(engine, 'project_user_quotas')
|
|
fake_quotas = {'id': 4,
|
|
'project_id': 'fake_project',
|
|
'user_id': 'fake_user',
|
|
'resource': 'instances',
|
|
'hard_limit': 10}
|
|
project_user_quotas.insert().execute(fake_quotas)
|
|
quota_usages = db_utils.get_table(engine, 'quota_usages')
|
|
reservations = db_utils.get_table(engine, 'reservations')
|
|
# Get the record
|
|
quota = project_user_quotas.select().execute().first()
|
|
quota_usage = quota_usages.select().execute().first()
|
|
reservation = reservations.select().execute().first()
|
|
|
|
self.assertEqual(quota['id'], 4)
|
|
self.assertEqual(quota['project_id'], 'fake_project')
|
|
self.assertEqual(quota['user_id'], 'fake_user')
|
|
self.assertEqual(quota['resource'], 'instances')
|
|
self.assertEqual(quota['hard_limit'], 10)
|
|
self.assertEqual(quota_usage['user_id'], None)
|
|
self.assertEqual(reservation['user_id'], None)
|
|
|
|
def _post_downgrade_203(self, engine):
|
|
try:
|
|
table_exist = True
|
|
db_utils.get_table(engine, 'project_user_quotas')
|
|
except Exception:
|
|
table_exist = False
|
|
quota_usages = db_utils.get_table(engine, 'quota_usages')
|
|
reservations = db_utils.get_table(engine, 'reservations')
|
|
|
|
# Get the record
|
|
quota_usage = quota_usages.select().execute().first()
|
|
reservation = reservations.select().execute().first()
|
|
|
|
self.assertFalse('user_id' in quota_usage)
|
|
self.assertFalse('user_id' in reservation)
|
|
self.assertFalse(table_exist)
|
|
|
|
def _check_204(self, engine, data):
|
|
if engine.name != 'sqlite':
|
|
return
|
|
|
|
meta = sqlalchemy.MetaData()
|
|
meta.bind = engine
|
|
reservations = sqlalchemy.Table('reservations', meta, autoload=True)
|
|
|
|
index_data = [(idx.name, idx.columns.keys())
|
|
for idx in reservations.indexes]
|
|
|
|
if engine.name == "postgresql":
|
|
# we can not get correct order of columns in index
|
|
# definition to postgresql using sqlalchemy. So we sort
|
|
# columns list before compare
|
|
# bug http://www.sqlalchemy.org/trac/ticket/2767
|
|
self.assertIn(
|
|
('reservations_uuid_idx', sorted(['uuid'])),
|
|
([(idx[0], sorted(idx[1])) for idx in index_data])
|
|
)
|
|
else:
|
|
self.assertIn(('reservations_uuid_idx', ['uuid']), index_data)
|
|
|
|
def _pre_upgrade_205(self, engine):
|
|
fake_instances = [dict(uuid='m205-uuid1', locked=True),
|
|
dict(uuid='m205-uuid2', locked=False)]
|
|
for table_name in ['instances', 'shadow_instances']:
|
|
table = db_utils.get_table(engine, table_name)
|
|
engine.execute(table.insert(), fake_instances)
|
|
|
|
def _check_205(self, engine, data):
|
|
for table_name in ['instances', 'shadow_instances']:
|
|
table = db_utils.get_table(engine, table_name)
|
|
rows = table.select().\
|
|
where(table.c.uuid.in_(['m205-uuid1', 'm205-uuid2'])).\
|
|
order_by(table.c.uuid).execute().fetchall()
|
|
self.assertEqual(rows[0]['locked_by'], 'admin')
|
|
self.assertEqual(rows[1]['locked_by'], None)
|
|
|
|
def _post_downgrade_205(self, engine):
|
|
for table_name in ['instances', 'shadow_instances']:
|
|
table = db_utils.get_table(engine, table_name)
|
|
rows = table.select().execute().fetchall()
|
|
self.assertFalse('locked_by' in rows[0])
|
|
|
|
def _pre_upgrade_206(self, engine):
|
|
instances = db_utils.get_table(engine, 'instances')
|
|
shadow_instances = db_utils.get_table(engine, 'shadow_instances')
|
|
|
|
data = [
|
|
{
|
|
'id': 650,
|
|
'deleted': 0,
|
|
},
|
|
{
|
|
'id': 651,
|
|
'deleted': 2,
|
|
},
|
|
]
|
|
for item in data:
|
|
instances.insert().values(item).execute()
|
|
shadow_instances.insert().values(item).execute()
|
|
return data
|
|
|
|
def _check_206(self, engine, data):
|
|
self.assertColumnExists(engine, 'instances', 'cleaned')
|
|
self.assertColumnExists(engine, 'shadow_instances', 'cleaned')
|
|
self.assertIndexMembers(engine, 'instances',
|
|
'instances_host_deleted_cleaned_idx',
|
|
['host', 'deleted', 'cleaned'])
|
|
|
|
instances = db_utils.get_table(engine, 'instances')
|
|
shadow_instances = db_utils.get_table(engine, 'shadow_instances')
|
|
|
|
def get_(table, ident):
|
|
return table.select().\
|
|
where(table.c.id == ident).\
|
|
execute().\
|
|
first()
|
|
|
|
for table in (instances, shadow_instances):
|
|
id_1 = get_(instances, 650)
|
|
self.assertEqual(0, id_1['deleted'])
|
|
self.assertEqual(0, id_1['cleaned'])
|
|
|
|
id_2 = get_(instances, 651)
|
|
self.assertEqual(2, id_2['deleted'])
|
|
self.assertEqual(1, id_2['cleaned'])
|
|
|
|
def _post_downgrade_206(self, engine):
|
|
self.assertColumnNotExists(engine, 'instances', 'cleaned')
|
|
self.assertColumnNotExists(engine, 'shadow_instances', 'cleaned')
|
|
|
|
def _207(self, engine, upgrade=False):
|
|
uniq_names = ['uniq_cell_name0deleted',
|
|
'uniq_cells0name0deleted']
|
|
if upgrade:
|
|
uniq_names = uniq_names[::-1]
|
|
cells = db_utils.get_table(engine, 'cells')
|
|
values = {'name': 'name',
|
|
'deleted': 0,
|
|
'transport_url': 'fake_transport_url'}
|
|
cells.insert().values(values).execute()
|
|
values['deleted'] = 1
|
|
cells.insert().values(values).execute()
|
|
values['deleted'] = 0
|
|
self.assertRaises(sqlalchemy.exc.IntegrityError,
|
|
cells.insert().execute,
|
|
values)
|
|
cells.delete().execute()
|
|
indexes = dict((i.name, i) for i in cells.indexes)
|
|
if indexes:
|
|
check_index_old = indexes.get(uniq_names[0])
|
|
check_index_new = indexes.get(uniq_names[1])
|
|
self.assertTrue(bool(check_index_old))
|
|
self.assertFalse(bool(check_index_new))
|
|
|
|
def _pre_upgrade_207(self, engine):
|
|
self._207(engine)
|
|
|
|
def _check_207(self, engine, data):
|
|
self._207(engine, upgrade=True)
|
|
|
|
def _post_downgrade_207(self, engine):
|
|
self._207(engine)
|
|
|
|
|
|
class TestBaremetalMigrations(BaseMigrationTestCase, CommonTestsMixIn):
|
|
"""Test sqlalchemy-migrate migrations."""
|
|
USER = "openstack_citest"
|
|
PASSWD = "openstack_citest"
|
|
DATABASE = "openstack_baremetal_citest"
|
|
|
|
def __init__(self, *args, **kwargs):
|
|
super(TestBaremetalMigrations, self).__init__(*args, **kwargs)
|
|
|
|
self.DEFAULT_CONFIG_FILE = os.path.join(os.path.dirname(__file__),
|
|
'../virt/baremetal/test_baremetal_migrations.conf')
|
|
# Test machines can set the NOVA_TEST_MIGRATIONS_CONF variable
|
|
# to override the location of the config file for migration testing
|
|
self.CONFIG_FILE_PATH = os.environ.get(
|
|
'BAREMETAL_TEST_MIGRATIONS_CONF',
|
|
self.DEFAULT_CONFIG_FILE)
|
|
self.MIGRATE_FILE = \
|
|
nova.virt.baremetal.db.sqlalchemy.migrate_repo.__file__
|
|
self.REPOSITORY = repository.Repository(
|
|
os.path.abspath(os.path.dirname(self.MIGRATE_FILE)))
|
|
|
|
def setUp(self):
|
|
super(TestBaremetalMigrations, self).setUp()
|
|
|
|
if self.migration is None:
|
|
self.migration = __import__('nova.virt.baremetal.db.migration',
|
|
globals(), locals(), ['INIT_VERSION'], -1)
|
|
self.INIT_VERSION = self.migration.INIT_VERSION
|
|
if self.migration_api is None:
|
|
temp = __import__('nova.virt.baremetal.db.sqlalchemy.migration',
|
|
globals(), locals(), ['versioning_api'], -1)
|
|
self.migration_api = temp.versioning_api
|
|
|
|
def _pre_upgrade_002(self, engine):
|
|
data = [{'id': 1, 'key': 'fake-key', 'image_path': '/dev/null',
|
|
'pxe_config_path': '/dev/null/', 'root_mb': 0, 'swap_mb': 0}]
|
|
table = db_utils.get_table(engine, 'bm_deployments')
|
|
engine.execute(table.insert(), data)
|
|
return data
|
|
|
|
def _check_002(self, engine, data):
|
|
self.assertRaises(sqlalchemy.exc.NoSuchTableError,
|
|
db_utils.get_table, engine, 'bm_deployments')
|
|
|
|
def _post_downgrade_004(self, engine):
|
|
bm_nodes = db_utils.get_table(engine, 'bm_nodes')
|
|
self.assertNotIn(u'instance_name', [c.name for c in bm_nodes.columns])
|
|
|
|
def _check_005(self, engine, data):
|
|
bm_nodes = db_utils.get_table(engine, 'bm_nodes')
|
|
columns = [c.name for c in bm_nodes.columns]
|
|
self.assertNotIn(u'prov_vlan_id', columns)
|
|
self.assertNotIn(u'registration_status', columns)
|
|
|
|
def _pre_upgrade_006(self, engine):
|
|
nodes = db_utils.get_table(engine, 'bm_nodes')
|
|
ifs = db_utils.get_table(engine, 'bm_interfaces')
|
|
# node 1 has two diffrent addresses in bm_nodes and bm_interfaces
|
|
engine.execute(nodes.insert(),
|
|
[{'id': 1,
|
|
'prov_mac_address': 'aa:aa:aa:aa:aa:aa'}])
|
|
engine.execute(ifs.insert(),
|
|
[{'id': 101,
|
|
'bm_node_id': 1,
|
|
'address': 'bb:bb:bb:bb:bb:bb'}])
|
|
# node 2 has one same address both in bm_nodes and bm_interfaces
|
|
engine.execute(nodes.insert(),
|
|
[{'id': 2,
|
|
'prov_mac_address': 'cc:cc:cc:cc:cc:cc'}])
|
|
engine.execute(ifs.insert(),
|
|
[{'id': 201,
|
|
'bm_node_id': 2,
|
|
'address': 'cc:cc:cc:cc:cc:cc'}])
|
|
|
|
def _check_006(self, engine, data):
|
|
ifs = db_utils.get_table(engine, 'bm_interfaces')
|
|
rows = ifs.select().\
|
|
where(ifs.c.bm_node_id == 1).\
|
|
execute().\
|
|
fetchall()
|
|
self.assertEqual(len(rows), 2)
|
|
rows = ifs.select().\
|
|
where(ifs.c.bm_node_id == 2).\
|
|
execute().\
|
|
fetchall()
|
|
self.assertEqual(len(rows), 1)
|
|
self.assertEqual(rows[0]['address'], 'cc:cc:cc:cc:cc:cc')
|
|
|
|
def _post_downgrade_006(self, engine):
|
|
ifs = db_utils.get_table(engine, 'bm_interfaces')
|
|
rows = ifs.select().where(ifs.c.bm_node_id == 1).execute().fetchall()
|
|
self.assertEqual(len(rows), 1)
|
|
self.assertEqual(rows[0]['address'], 'bb:bb:bb:bb:bb:bb')
|
|
|
|
rows = ifs.select().where(ifs.c.bm_node_id == 2).execute().fetchall()
|
|
self.assertEqual(len(rows), 0)
|
|
|
|
def _check_007(self, engine, data):
|
|
bm_nodes = db_utils.get_table(engine, 'bm_nodes')
|
|
columns = [c.name for c in bm_nodes.columns]
|
|
self.assertNotIn(u'prov_mac_address', columns)
|
|
|
|
def _check_008(self, engine, data):
|
|
self.assertRaises(sqlalchemy.exc.NoSuchTableError,
|
|
db_utils.get_table, engine, 'bm_pxe_ips')
|
|
|
|
def _post_downgrade_008(self, engine):
|
|
db_utils.get_table(engine, 'bm_pxe_ips')
|
|
|
|
|
|
class ProjectTestCase(test.TestCase):
|
|
|
|
def test_all_migrations_have_downgrade(self):
|
|
topdir = os.path.normpath(os.path.dirname(__file__) + '/../../../')
|
|
py_glob = os.path.join(topdir, "nova", "db", "sqlalchemy",
|
|
"migrate_repo", "versions", "*.py")
|
|
|
|
missing_downgrade = []
|
|
for path in glob.iglob(py_glob):
|
|
has_upgrade = False
|
|
has_downgrade = False
|
|
with open(path, "r") as f:
|
|
for line in f:
|
|
if 'def upgrade(' in line:
|
|
has_upgrade = True
|
|
if 'def downgrade(' in line:
|
|
has_downgrade = True
|
|
|
|
if has_upgrade and not has_downgrade:
|
|
fname = os.path.basename(path)
|
|
missing_downgrade.append(fname)
|
|
|
|
helpful_msg = (_("The following migrations are missing a downgrade:"
|
|
"\n\t%s") % '\n\t'.join(sorted(missing_downgrade)))
|
|
self.assert_(not missing_downgrade, helpful_msg)
|