keystone/keystone/common/sql/migrations/env.py
Stephen Finucane 2bf70a10a2 sql: Fix incorrect columns
In these instances, we take the migrations to be the "official" version
- since they're stricter in almost all cases - updating the models to
suit.

This change highlights a slight issue in our use of a config option in
our database schema, which we shouldn't really do. A TODO is left to
address this later. We can also remove a now-unnecessary TODO from our
initial migration related to the same issue: we have our own tooling for
migrations that *does* load and register config options so there is no
longer an issue here.

Change-Id: I906cb8f7b76833c880a40c1aa0584fe7ab93cb7a
Signed-off-by: Stephen Finucane <sfinucan@redhat.com>
2023-07-03 12:32:58 +01:00

196 lines
6.4 KiB
Python

# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from logging.config import fileConfig
from alembic import context
from sqlalchemy import engine_from_config
from sqlalchemy import pool
from keystone.common.sql import core
from keystone.common.sql.migrations import autogen
# this is the Alembic Config object, which provides
# access to the values within the .ini file in use.
config = context.config
# interpret the config file for Python logging unless we're told not to;
# this line sets up loggers basically.
if config.attributes.get('configure_logger', True):
fileConfig(config.config_file_name)
# keystone model MetaData object
target_metadata = core.ModelBase.metadata
def include_object(object, name, type_, reflected, compare_to):
BORKED_COLUMNS = (
)
BORKED_UNIQUE_CONSTRAINTS = (
)
BORKED_FK_CONSTRAINTS = (
# removed fks
('application_credential_access_rule', ['access_rule_id']),
('limit', ['registered_limit_id']),
('registered_limit', ['service_id']),
('registered_limit', ['region_id']),
('endpoint', ['region_id']),
# added fks
('application_credential_access_rule', ['access_rule_id']),
('endpoint', ['region_id']),
('assignment', ['role_id']),
)
BORKED_INDEXES = (
# removed indexes
('access_rule', ['external_id']),
('access_rule', ['user_id']),
('revocation_event', ['revoked_at']),
('system_assignment', ['actor_id']),
('user', ['default_project_id']),
# added indexes
('access_rule', ['external_id']),
('access_rule', ['user_id']),
('access_token', ['consumer_id']),
('endpoint', ['service_id']),
('revocation_event', ['revoked_at']),
('user', ['default_project_id']),
('user_group_membership', ['group_id']),
(
'trust',
[
'trustor_user_id',
'trustee_user_id',
'project_id',
'impersonation',
'expires_at',
'expires_at_int',
],
),
)
# NOTE(stephenfin): By skipping these items, we skip *all* changes to the
# affected item. However, we only want to skip the actual things we know
# about untl we have enough time to fix them. These issues are listed in
# keystone.tests.unit.common.sql.test_upgrades.KeystoneModelsMigrationsSync
# However, this isn't an issue since the test is more specific and will
# catch other issues and anyone making changes to the columns and hoping to
# autogenerate them would need to fix the latent issue first anyway.
if type_ == 'column':
return (object.table.name, name) not in BORKED_COLUMNS
if type_ == 'unique_constraint':
columns = [c.name for c in object.columns]
return (object.table.name, columns) not in BORKED_UNIQUE_CONSTRAINTS
if type_ == 'foreign_key_constraint':
columns = [c.name for c in object.columns]
return (object.table.name, columns) not in BORKED_FK_CONSTRAINTS
if type_ == 'index':
columns = [c.name for c in object.columns]
return (object.table.name, columns) not in BORKED_INDEXES
return True
def include_name(name, type_, parent_names):
"""Determine which tables or columns to skip.
This is used where we have migrations that are out-of-sync with the models.
"""
REMOVED_TABLES = ('token',)
if type_ == 'table':
return name not in REMOVED_TABLES
return True
def run_migrations_offline():
"""Run migrations in 'offline' mode.
This configures the context with just a URL and not an Engine, though an
Engine is acceptable here as well. By skipping the Engine creation we
don't even need a DBAPI to be available.
Calls to context.execute() here emit the given string to the script output.
"""
url = config.get_main_option("sqlalchemy.url")
context.configure(
url=url,
target_metadata=target_metadata,
render_as_batch=True,
include_name=include_name,
include_object=include_object,
literal_binds=True,
dialect_opts={"paramstyle": "named"},
)
with context.begin_transaction():
context.run_migrations()
def run_migrations_online():
"""Run migrations in 'online' mode.
In this scenario we need to create an Engine and associate a connection
with the context.
This is modified from the default based on the below, since we want to
share an engine when unit testing so in-memory database testing actually
works.
https://alembic.sqlalchemy.org/en/latest/cookbook.html#connection-sharing
"""
connectable = config.attributes.get('connection', None)
if connectable is None:
# only create Engine if we don't have a Connection from the outside
connectable = engine_from_config(
config.get_section(config.config_ini_section),
prefix="sqlalchemy.",
poolclass=pool.NullPool,
)
with connectable.connect() as connection:
context.configure(
connection=connection,
target_metadata=target_metadata,
render_as_batch=True,
include_name=include_name,
include_object=include_object,
process_revision_directives=autogen.process_revision_directives, # noqa: E501
)
with context.begin_transaction():
context.run_migrations()
else:
context.configure(
connection=connectable,
target_metadata=target_metadata,
render_as_batch=True,
include_name=include_name,
include_object=include_object,
process_revision_directives=autogen.process_revision_directives,
)
with context.begin_transaction():
context.run_migrations()
if context.is_offline_mode():
run_migrations_offline()
else:
run_migrations_online()