Your IP : 18.116.63.107
# testing/suite/test_reflection.py
# Copyright (C) 2005-2024 the SQLAlchemy authors and contributors
# <see AUTHORS file>
#
# This module is part of SQLAlchemy and is released under
# the MIT License: https://www.opensource.org/licenses/mit-license.php
# mypy: ignore-errors
import operator
import re
import sqlalchemy as sa
from .. import config
from .. import engines
from .. import eq_
from .. import expect_raises
from .. import expect_raises_message
from .. import expect_warnings
from .. import fixtures
from .. import is_
from ..provision import get_temp_table_name
from ..provision import temp_table_keyword_args
from ..schema import Column
from ..schema import Table
from ... import event
from ... import ForeignKey
from ... import func
from ... import Identity
from ... import inspect
from ... import Integer
from ... import MetaData
from ... import String
from ... import testing
from ... import types as sql_types
from ...engine import Inspector
from ...engine import ObjectKind
from ...engine import ObjectScope
from ...exc import NoSuchTableError
from ...exc import UnreflectableTableError
from ...schema import DDL
from ...schema import Index
from ...sql.elements import quoted_name
from ...sql.schema import BLANK_SCHEMA
from ...testing import ComparesIndexes
from ...testing import ComparesTables
from ...testing import is_false
from ...testing import is_true
from ...testing import mock
metadata, users = None, None
class OneConnectionTablesTest(fixtures.TablesTest):
@classmethod
def setup_bind(cls):
# TODO: when temp tables are subject to server reset,
# this will also have to disable that server reset from
# happening
if config.requirements.independent_connections.enabled:
from sqlalchemy import pool
return engines.testing_engine(
options=dict(poolclass=pool.StaticPool, scope="class"),
)
else:
return config.db
class HasTableTest(OneConnectionTablesTest):
__backend__ = True
@classmethod
def define_tables(cls, metadata):
Table(
"test_table",
metadata,
Column("id", Integer, primary_key=True),
Column("data", String(50)),
)
if testing.requires.schemas.enabled:
Table(
"test_table_s",
metadata,
Column("id", Integer, primary_key=True),
Column("data", String(50)),
schema=config.test_schema,
)
if testing.requires.view_reflection:
cls.define_views(metadata)
if testing.requires.has_temp_table.enabled:
cls.define_temp_tables(metadata)
@classmethod
def define_views(cls, metadata):
query = "CREATE VIEW vv AS SELECT id, data FROM test_table"
event.listen(metadata, "after_create", DDL(query))
event.listen(metadata, "before_drop", DDL("DROP VIEW vv"))
if testing.requires.schemas.enabled:
query = (
"CREATE VIEW %s.vv AS SELECT id, data FROM %s.test_table_s"
% (
config.test_schema,
config.test_schema,
)
)
event.listen(metadata, "after_create", DDL(query))
event.listen(
metadata,
"before_drop",
DDL("DROP VIEW %s.vv" % (config.test_schema)),
)
@classmethod
def temp_table_name(cls):
return get_temp_table_name(
config, config.db, f"user_tmp_{config.ident}"
)
@classmethod
def define_temp_tables(cls, metadata):
kw = temp_table_keyword_args(config, config.db)
table_name = cls.temp_table_name()
user_tmp = Table(
table_name,
metadata,
Column("id", sa.INT, primary_key=True),
Column("name", sa.VARCHAR(50)),
**kw,
)
if (
testing.requires.view_reflection.enabled
and testing.requires.temporary_views.enabled
):
event.listen(
user_tmp,
"after_create",
DDL(
"create temporary view user_tmp_v as "
"select * from user_tmp_%s" % config.ident
),
)
event.listen(user_tmp, "before_drop", DDL("drop view user_tmp_v"))
def test_has_table(self):
with config.db.begin() as conn:
is_true(config.db.dialect.has_table(conn, "test_table"))
is_false(config.db.dialect.has_table(conn, "test_table_s"))
is_false(config.db.dialect.has_table(conn, "nonexistent_table"))
def test_has_table_cache(self, metadata):
insp = inspect(config.db)
is_true(insp.has_table("test_table"))
nt = Table("new_table", metadata, Column("col", Integer))
is_false(insp.has_table("new_table"))
nt.create(config.db)
try:
is_false(insp.has_table("new_table"))
insp.clear_cache()
is_true(insp.has_table("new_table"))
finally:
nt.drop(config.db)
@testing.requires.schemas
def test_has_table_schema(self):
with config.db.begin() as conn:
is_false(
config.db.dialect.has_table(
conn, "test_table", schema=config.test_schema
)
)
is_true(
config.db.dialect.has_table(
conn, "test_table_s", schema=config.test_schema
)
)
is_false(
config.db.dialect.has_table(
conn, "nonexistent_table", schema=config.test_schema
)
)
@testing.requires.schemas
def test_has_table_nonexistent_schema(self):
with config.db.begin() as conn:
is_false(
config.db.dialect.has_table(
conn, "test_table", schema="nonexistent_schema"
)
)
@testing.requires.views
def test_has_table_view(self, connection):
insp = inspect(connection)
is_true(insp.has_table("vv"))
@testing.requires.has_temp_table
def test_has_table_temp_table(self, connection):
insp = inspect(connection)
temp_table_name = self.temp_table_name()
is_true(insp.has_table(temp_table_name))
@testing.requires.has_temp_table
@testing.requires.view_reflection
@testing.requires.temporary_views
def test_has_table_temp_view(self, connection):
insp = inspect(connection)
is_true(insp.has_table("user_tmp_v"))
@testing.requires.views
@testing.requires.schemas
def test_has_table_view_schema(self, connection):
insp = inspect(connection)
is_true(insp.has_table("vv", config.test_schema))
class HasIndexTest(fixtures.TablesTest):
__backend__ = True
@classmethod
def define_tables(cls, metadata):
tt = Table(
"test_table",
metadata,
Column("id", Integer, primary_key=True),
Column("data", String(50)),
Column("data2", String(50)),
)
Index("my_idx", tt.c.data)
if testing.requires.schemas.enabled:
tt = Table(
"test_table",
metadata,
Column("id", Integer, primary_key=True),
Column("data", String(50)),
schema=config.test_schema,
)
Index("my_idx_s", tt.c.data)
kind = testing.combinations("dialect", "inspector", argnames="kind")
def _has_index(self, kind, conn):
if kind == "dialect":
return lambda *a, **k: config.db.dialect.has_index(conn, *a, **k)
else:
return inspect(conn).has_index
@kind
def test_has_index(self, kind, connection, metadata):
meth = self._has_index(kind, connection)
assert meth("test_table", "my_idx")
assert not meth("test_table", "my_idx_s")
assert not meth("nonexistent_table", "my_idx")
assert not meth("test_table", "nonexistent_idx")
assert not meth("test_table", "my_idx_2")
assert not meth("test_table_2", "my_idx_3")
idx = Index("my_idx_2", self.tables.test_table.c.data2)
tbl = Table(
"test_table_2",
metadata,
Column("foo", Integer),
Index("my_idx_3", "foo"),
)
idx.create(connection)
tbl.create(connection)
try:
if kind == "inspector":
assert not meth("test_table", "my_idx_2")
assert not meth("test_table_2", "my_idx_3")
meth.__self__.clear_cache()
assert meth("test_table", "my_idx_2") is True
assert meth("test_table_2", "my_idx_3") is True
finally:
tbl.drop(connection)
idx.drop(connection)
@testing.requires.schemas
@kind
def test_has_index_schema(self, kind, connection):
meth = self._has_index(kind, connection)
assert meth("test_table", "my_idx_s", schema=config.test_schema)
assert not meth("test_table", "my_idx", schema=config.test_schema)
assert not meth(
"nonexistent_table", "my_idx_s", schema=config.test_schema
)
assert not meth(
"test_table", "nonexistent_idx_s", schema=config.test_schema
)
class BizarroCharacterFKResolutionTest(fixtures.TestBase):
"""tests for #10275"""
__backend__ = True
@testing.combinations(
("id",), ("(3)",), ("col%p",), ("[brack]",), argnames="columnname"
)
@testing.variation("use_composite", [True, False])
@testing.combinations(
("plain",),
("(2)",),
("per % cent",),
("[brackets]",),
argnames="tablename",
)
def test_fk_ref(
self, connection, metadata, use_composite, tablename, columnname
):
tt = Table(
tablename,
metadata,
Column(columnname, Integer, key="id", primary_key=True),
test_needs_fk=True,
)
if use_composite:
tt.append_column(Column("id2", Integer, primary_key=True))
if use_composite:
Table(
"other",
metadata,
Column("id", Integer, primary_key=True),
Column("ref", Integer),
Column("ref2", Integer),
sa.ForeignKeyConstraint(["ref", "ref2"], [tt.c.id, tt.c.id2]),
test_needs_fk=True,
)
else:
Table(
"other",
metadata,
Column("id", Integer, primary_key=True),
Column("ref", ForeignKey(tt.c.id)),
test_needs_fk=True,
)
metadata.create_all(connection)
m2 = MetaData()
o2 = Table("other", m2, autoload_with=connection)
t1 = m2.tables[tablename]
assert o2.c.ref.references(t1.c[0])
if use_composite:
assert o2.c.ref2.references(t1.c[1])
class QuotedNameArgumentTest(fixtures.TablesTest):
run_create_tables = "once"
__backend__ = True
@classmethod
def define_tables(cls, metadata):
Table(
"quote ' one",
metadata,
Column("id", Integer),
Column("name", String(50)),
Column("data", String(50)),
Column("related_id", Integer),
sa.PrimaryKeyConstraint("id", name="pk quote ' one"),
sa.Index("ix quote ' one", "name"),
sa.UniqueConstraint(
"data",
name="uq quote' one",
),
sa.ForeignKeyConstraint(
["id"], ["related.id"], name="fk quote ' one"
),
sa.CheckConstraint("name != 'foo'", name="ck quote ' one"),
comment=r"""quote ' one comment""",
test_needs_fk=True,
)
if testing.requires.symbol_names_w_double_quote.enabled:
Table(
'quote " two',
metadata,
Column("id", Integer),
Column("name", String(50)),
Column("data", String(50)),
Column("related_id", Integer),
sa.PrimaryKeyConstraint("id", name='pk quote " two'),
sa.Index('ix quote " two', "name"),
sa.UniqueConstraint(
"data",
name='uq quote" two',
),
sa.ForeignKeyConstraint(
["id"], ["related.id"], name='fk quote " two'
),
sa.CheckConstraint("name != 'foo'", name='ck quote " two '),
comment=r"""quote " two comment""",
test_needs_fk=True,
)
Table(
"related",
metadata,
Column("id", Integer, primary_key=True),
Column("related", Integer),
test_needs_fk=True,
)
if testing.requires.view_column_reflection.enabled:
if testing.requires.symbol_names_w_double_quote.enabled:
names = [
"quote ' one",
'quote " two',
]
else:
names = [
"quote ' one",
]
for name in names:
query = "CREATE VIEW %s AS SELECT * FROM %s" % (
config.db.dialect.identifier_preparer.quote(
"view %s" % name
),
config.db.dialect.identifier_preparer.quote(name),
)
event.listen(metadata, "after_create", DDL(query))
event.listen(
metadata,
"before_drop",
DDL(
"DROP VIEW %s"
% config.db.dialect.identifier_preparer.quote(
"view %s" % name
)
),
)
def quote_fixtures(fn):
return testing.combinations(
("quote ' one",),
('quote " two', testing.requires.symbol_names_w_double_quote),
)(fn)
@quote_fixtures
def test_get_table_options(self, name):
insp = inspect(config.db)
if testing.requires.reflect_table_options.enabled:
res = insp.get_table_options(name)
is_true(isinstance(res, dict))
else:
with expect_raises(NotImplementedError):
res = insp.get_table_options(name)
@quote_fixtures
@testing.requires.view_column_reflection
def test_get_view_definition(self, name):
insp = inspect(config.db)
assert insp.get_view_definition("view %s" % name)
@quote_fixtures
def test_get_columns(self, name):
insp = inspect(config.db)
assert insp.get_columns(name)
@quote_fixtures
def test_get_pk_constraint(self, name):
insp = inspect(config.db)
assert insp.get_pk_constraint(name)
@quote_fixtures
def test_get_foreign_keys(self, name):
insp = inspect(config.db)
assert insp.get_foreign_keys(name)
@quote_fixtures
def test_get_indexes(self, name):
insp = inspect(config.db)
assert insp.get_indexes(name)
@quote_fixtures
@testing.requires.unique_constraint_reflection
def test_get_unique_constraints(self, name):
insp = inspect(config.db)
assert insp.get_unique_constraints(name)
@quote_fixtures
@testing.requires.comment_reflection
def test_get_table_comment(self, name):
insp = inspect(config.db)
assert insp.get_table_comment(name)
@quote_fixtures
@testing.requires.check_constraint_reflection
def test_get_check_constraints(self, name):
insp = inspect(config.db)
assert insp.get_check_constraints(name)
def _multi_combination(fn):
schema = testing.combinations(
None,
(
lambda: config.test_schema,
testing.requires.schemas,
),
argnames="schema",
)
scope = testing.combinations(
ObjectScope.DEFAULT,
ObjectScope.TEMPORARY,
ObjectScope.ANY,
argnames="scope",
)
kind = testing.combinations(
ObjectKind.TABLE,
ObjectKind.VIEW,
ObjectKind.MATERIALIZED_VIEW,
ObjectKind.ANY,
ObjectKind.ANY_VIEW,
ObjectKind.TABLE | ObjectKind.VIEW,
ObjectKind.TABLE | ObjectKind.MATERIALIZED_VIEW,
argnames="kind",
)
filter_names = testing.combinations(True, False, argnames="use_filter")
return schema(scope(kind(filter_names(fn))))
class ComponentReflectionTest(ComparesTables, OneConnectionTablesTest):
run_inserts = run_deletes = None
__backend__ = True
@classmethod
def define_tables(cls, metadata):
cls.define_reflected_tables(metadata, None)
if testing.requires.schemas.enabled:
cls.define_reflected_tables(metadata, testing.config.test_schema)
@classmethod
def define_reflected_tables(cls, metadata, schema):
if schema:
schema_prefix = schema + "."
else:
schema_prefix = ""
if testing.requires.self_referential_foreign_keys.enabled:
parent_id_args = (
ForeignKey(
"%susers.user_id" % schema_prefix, name="user_id_fk"
),
)
else:
parent_id_args = ()
users = Table(
"users",
metadata,
Column("user_id", sa.INT, primary_key=True),
Column("test1", sa.CHAR(5), nullable=False),
Column("test2", sa.Float(), nullable=False),
Column("parent_user_id", sa.Integer, *parent_id_args),
sa.CheckConstraint(
"test2 > 0",
name="zz_test2_gt_zero",
comment="users check constraint",
),
sa.CheckConstraint("test2 <= 1000"),
schema=schema,
test_needs_fk=True,
)
Table(
"dingalings",
metadata,
Column("dingaling_id", sa.Integer, primary_key=True),
Column(
"address_id",
sa.Integer,
ForeignKey(
"%semail_addresses.address_id" % schema_prefix,
name="zz_email_add_id_fg",
comment="di fk comment",
),
),
Column(
"id_user",
sa.Integer,
ForeignKey("%susers.user_id" % schema_prefix),
),
Column("data", sa.String(30), unique=True),
sa.CheckConstraint(
"address_id > 0 AND address_id < 1000",
name="address_id_gt_zero",
),
sa.UniqueConstraint(
"address_id",
"dingaling_id",
name="zz_dingalings_multiple",
comment="di unique comment",
),
schema=schema,
test_needs_fk=True,
)
Table(
"email_addresses",
metadata,
Column("address_id", sa.Integer),
Column("remote_user_id", sa.Integer, ForeignKey(users.c.user_id)),
Column("email_address", sa.String(20), index=True),
sa.PrimaryKeyConstraint(
"address_id", name="email_ad_pk", comment="ea pk comment"
),
schema=schema,
test_needs_fk=True,
)
Table(
"comment_test",
metadata,
Column("id", sa.Integer, primary_key=True, comment="id comment"),
Column("data", sa.String(20), comment="data % comment"),
Column(
"d2",
sa.String(20),
comment=r"""Comment types type speedily ' " \ '' Fun!""",
),
Column("d3", sa.String(42), comment="Comment\nwith\rescapes"),
schema=schema,
comment=r"""the test % ' " \ table comment""",
)
Table(
"no_constraints",
metadata,
Column("data", sa.String(20)),
schema=schema,
comment="no\nconstraints\rhas\fescaped\vcomment",
)
if testing.requires.cross_schema_fk_reflection.enabled:
if schema is None:
Table(
"local_table",
metadata,
Column("id", sa.Integer, primary_key=True),
Column("data", sa.String(20)),
Column(
"remote_id",
ForeignKey(
"%s.remote_table_2.id" % testing.config.test_schema
),
),
test_needs_fk=True,
schema=config.db.dialect.default_schema_name,
)
else:
Table(
"remote_table",
metadata,
Column("id", sa.Integer, primary_key=True),
Column(
"local_id",
ForeignKey(
"%s.local_table.id"
% config.db.dialect.default_schema_name
),
),
Column("data", sa.String(20)),
schema=schema,
test_needs_fk=True,
)
Table(
"remote_table_2",
metadata,
Column("id", sa.Integer, primary_key=True),
Column("data", sa.String(20)),
schema=schema,
test_needs_fk=True,
)
if testing.requires.index_reflection.enabled:
Index("users_t_idx", users.c.test1, users.c.test2, unique=True)
Index(
"users_all_idx", users.c.user_id, users.c.test2, users.c.test1
)
if not schema:
# test_needs_fk is at the moment to force MySQL InnoDB
noncol_idx_test_nopk = Table(
"noncol_idx_test_nopk",
metadata,
Column("q", sa.String(5)),
test_needs_fk=True,
)
noncol_idx_test_pk = Table(
"noncol_idx_test_pk",
metadata,
Column("id", sa.Integer, primary_key=True),
Column("q", sa.String(5)),
test_needs_fk=True,
)
if (
testing.requires.indexes_with_ascdesc.enabled
and testing.requires.reflect_indexes_with_ascdesc.enabled
):
Index("noncol_idx_nopk", noncol_idx_test_nopk.c.q.desc())
Index("noncol_idx_pk", noncol_idx_test_pk.c.q.desc())
if testing.requires.view_column_reflection.enabled:
cls.define_views(metadata, schema)
if not schema and testing.requires.temp_table_reflection.enabled:
cls.define_temp_tables(metadata)
@classmethod
def temp_table_name(cls):
return get_temp_table_name(
config, config.db, f"user_tmp_{config.ident}"
)
@classmethod
def define_temp_tables(cls, metadata):
kw = temp_table_keyword_args(config, config.db)
table_name = cls.temp_table_name()
user_tmp = Table(
table_name,
metadata,
Column("id", sa.INT, primary_key=True),
Column("name", sa.VARCHAR(50)),
Column("foo", sa.INT),
# disambiguate temp table unique constraint names. this is
# pretty arbitrary for a generic dialect however we are doing
# it to suit SQL Server which will produce name conflicts for
# unique constraints created against temp tables in different
# databases.
# https://www.arbinada.com/en/node/1645
sa.UniqueConstraint("name", name=f"user_tmp_uq_{config.ident}"),
sa.Index("user_tmp_ix", "foo"),
**kw,
)
if (
testing.requires.view_reflection.enabled
and testing.requires.temporary_views.enabled
):
event.listen(
user_tmp,
"after_create",
DDL(
"create temporary view user_tmp_v as "
"select * from user_tmp_%s" % config.ident
),
)
event.listen(user_tmp, "before_drop", DDL("drop view user_tmp_v"))
@classmethod
def define_views(cls, metadata, schema):
if testing.requires.materialized_views.enabled:
materialized = {"dingalings"}
else:
materialized = set()
for table_name in ("users", "email_addresses", "dingalings"):
fullname = table_name
if schema:
fullname = f"{schema}.{table_name}"
view_name = fullname + "_v"
prefix = "MATERIALIZED " if table_name in materialized else ""
query = (
f"CREATE {prefix}VIEW {view_name} AS SELECT * FROM {fullname}"
)
event.listen(metadata, "after_create", DDL(query))
if table_name in materialized:
index_name = "mat_index"
if schema and testing.against("oracle"):
index_name = f"{schema}.{index_name}"
idx = f"CREATE INDEX {index_name} ON {view_name}(data)"
event.listen(metadata, "after_create", DDL(idx))
event.listen(
metadata, "before_drop", DDL(f"DROP {prefix}VIEW {view_name}")
)
def _resolve_kind(self, kind, tables, views, materialized):
res = {}
if ObjectKind.TABLE in kind:
res.update(tables)
if ObjectKind.VIEW in kind:
res.update(views)
if ObjectKind.MATERIALIZED_VIEW in kind:
res.update(materialized)
return res
def _resolve_views(self, views, materialized):
if not testing.requires.view_column_reflection.enabled:
materialized.clear()
views.clear()
elif not testing.requires.materialized_views.enabled:
views.update(materialized)
materialized.clear()
def _resolve_names(self, schema, scope, filter_names, values):
scope_filter = lambda _: True # noqa: E731
if scope is ObjectScope.DEFAULT:
scope_filter = lambda k: "tmp" not in k[1] # noqa: E731
if scope is ObjectScope.TEMPORARY:
scope_filter = lambda k: "tmp" in k[1] # noqa: E731
removed = {
None: {"remote_table", "remote_table_2"},
testing.config.test_schema: {
"local_table",
"noncol_idx_test_nopk",
"noncol_idx_test_pk",
"user_tmp_v",
self.temp_table_name(),
},
}
if not testing.requires.cross_schema_fk_reflection.enabled:
removed[None].add("local_table")
removed[testing.config.test_schema].update(
["remote_table", "remote_table_2"]
)
if not testing.requires.index_reflection.enabled:
removed[None].update(
["noncol_idx_test_nopk", "noncol_idx_test_pk"]
)
if (
not testing.requires.temp_table_reflection.enabled
or not testing.requires.temp_table_names.enabled
):
removed[None].update(["user_tmp_v", self.temp_table_name()])
if not testing.requires.temporary_views.enabled:
removed[None].update(["user_tmp_v"])
res = {
k: v
for k, v in values.items()
if scope_filter(k)
and k[1] not in removed[schema]
and (not filter_names or k[1] in filter_names)
}
return res
def exp_options(
self,
schema=None,
scope=ObjectScope.ANY,
kind=ObjectKind.ANY,
filter_names=None,
):
materialized = {(schema, "dingalings_v"): mock.ANY}
views = {
(schema, "email_addresses_v"): mock.ANY,
(schema, "users_v"): mock.ANY,
(schema, "user_tmp_v"): mock.ANY,
}
self._resolve_views(views, materialized)
tables = {
(schema, "users"): mock.ANY,
(schema, "dingalings"): mock.ANY,
(schema, "email_addresses"): mock.ANY,
(schema, "comment_test"): mock.ANY,
(schema, "no_constraints"): mock.ANY,
(schema, "local_table"): mock.ANY,
(schema, "remote_table"): mock.ANY,
(schema, "remote_table_2"): mock.ANY,
(schema, "noncol_idx_test_nopk"): mock.ANY,
(schema, "noncol_idx_test_pk"): mock.ANY,
(schema, self.temp_table_name()): mock.ANY,
}
res = self._resolve_kind(kind, tables, views, materialized)
res = self._resolve_names(schema, scope, filter_names, res)
return res
def exp_comments(
self,
schema=None,
scope=ObjectScope.ANY,
kind=ObjectKind.ANY,
filter_names=None,
):
empty = {"text": None}
materialized = {(schema, "dingalings_v"): empty}
views = {
(schema, "email_addresses_v"): empty,
(schema, "users_v"): empty,
(schema, "user_tmp_v"): empty,
}
self._resolve_views(views, materialized)
tables = {
(schema, "users"): empty,
(schema, "dingalings"): empty,
(schema, "email_addresses"): empty,
(schema, "comment_test"): {
"text": r"""the test % ' " \ table comment"""
},
(schema, "no_constraints"): {
"text": "no\nconstraints\rhas\fescaped\vcomment"
},
(schema, "local_table"): empty,
(schema, "remote_table"): empty,
(schema, "remote_table_2"): empty,
(schema, "noncol_idx_test_nopk"): empty,
(schema, "noncol_idx_test_pk"): empty,
(schema, self.temp_table_name()): empty,
}
res = self._resolve_kind(kind, tables, views, materialized)
res = self._resolve_names(schema, scope, filter_names, res)
return res
def exp_columns(
self,
schema=None,
scope=ObjectScope.ANY,
kind=ObjectKind.ANY,
filter_names=None,
):
def col(
name, auto=False, default=mock.ANY, comment=None, nullable=True
):
res = {
"name": name,
"autoincrement": auto,
"type": mock.ANY,
"default": default,
"comment": comment,
"nullable": nullable,
}
if auto == "omit":
res.pop("autoincrement")
return res
def pk(name, **kw):
kw = {"auto": True, "default": mock.ANY, "nullable": False, **kw}
return col(name, **kw)
materialized = {
(schema, "dingalings_v"): [
col("dingaling_id", auto="omit", nullable=mock.ANY),
col("address_id"),
col("id_user"),
col("data"),
]
}
views = {
(schema, "email_addresses_v"): [
col("address_id", auto="omit", nullable=mock.ANY),
col("remote_user_id"),
col("email_address"),
],
(schema, "users_v"): [
col("user_id", auto="omit", nullable=mock.ANY),
col("test1", nullable=mock.ANY),
col("test2", nullable=mock.ANY),
col("parent_user_id"),
],
(schema, "user_tmp_v"): [
col("id", auto="omit", nullable=mock.ANY),
col("name"),
col("foo"),
],
}
self._resolve_views(views, materialized)
tables = {
(schema, "users"): [
pk("user_id"),
col("test1", nullable=False),
col("test2", nullable=False),
col("parent_user_id"),
],
(schema, "dingalings"): [
pk("dingaling_id"),
col("address_id"),
col("id_user"),
col("data"),
],
(schema, "email_addresses"): [
pk("address_id"),
col("remote_user_id"),
col("email_address"),
],
(schema, "comment_test"): [
pk("id", comment="id comment"),
col("data", comment="data % comment"),
col(
"d2",
comment=r"""Comment types type speedily ' " \ '' Fun!""",
),
col("d3", comment="Comment\nwith\rescapes"),
],
(schema, "no_constraints"): [col("data")],
(schema, "local_table"): [pk("id"), col("data"), col("remote_id")],
(schema, "remote_table"): [pk("id"), col("local_id"), col("data")],
(schema, "remote_table_2"): [pk("id"), col("data")],
(schema, "noncol_idx_test_nopk"): [col("q")],
(schema, "noncol_idx_test_pk"): [pk("id"), col("q")],
(schema, self.temp_table_name()): [
pk("id"),
col("name"),
col("foo"),
],
}
res = self._resolve_kind(kind, tables, views, materialized)
res = self._resolve_names(schema, scope, filter_names, res)
return res
@property
def _required_column_keys(self):
return {"name", "type", "nullable", "default"}
def exp_pks(
self,
schema=None,
scope=ObjectScope.ANY,
kind=ObjectKind.ANY,
filter_names=None,
):
def pk(*cols, name=mock.ANY, comment=None):
return {
"constrained_columns": list(cols),
"name": name,
"comment": comment,
}
empty = pk(name=None)
if testing.requires.materialized_views_reflect_pk.enabled:
materialized = {(schema, "dingalings_v"): pk("dingaling_id")}
else:
materialized = {(schema, "dingalings_v"): empty}
views = {
(schema, "email_addresses_v"): empty,
(schema, "users_v"): empty,
(schema, "user_tmp_v"): empty,
}
self._resolve_views(views, materialized)
tables = {
(schema, "users"): pk("user_id"),
(schema, "dingalings"): pk("dingaling_id"),
(schema, "email_addresses"): pk(
"address_id", name="email_ad_pk", comment="ea pk comment"
),
(schema, "comment_test"): pk("id"),
(schema, "no_constraints"): empty,
(schema, "local_table"): pk("id"),
(schema, "remote_table"): pk("id"),
(schema, "remote_table_2"): pk("id"),
(schema, "noncol_idx_test_nopk"): empty,
(schema, "noncol_idx_test_pk"): pk("id"),
(schema, self.temp_table_name()): pk("id"),
}
if not testing.requires.reflects_pk_names.enabled:
for val in tables.values():
if val["name"] is not None:
val["name"] = mock.ANY
res = self._resolve_kind(kind, tables, views, materialized)
res = self._resolve_names(schema, scope, filter_names, res)
return res
@property
def _required_pk_keys(self):
return {"name", "constrained_columns"}
def exp_fks(
self,
schema=None,
scope=ObjectScope.ANY,
kind=ObjectKind.ANY,
filter_names=None,
):
class tt:
def __eq__(self, other):
return (
other is None
or config.db.dialect.default_schema_name == other
)
def fk(
cols,
ref_col,
ref_table,
ref_schema=schema,
name=mock.ANY,
comment=None,
):
return {
"constrained_columns": cols,
"referred_columns": ref_col,
"name": name,
"options": mock.ANY,
"referred_schema": (
ref_schema if ref_schema is not None else tt()
),
"referred_table": ref_table,
"comment": comment,
}
materialized = {(schema, "dingalings_v"): []}
views = {
(schema, "email_addresses_v"): [],
(schema, "users_v"): [],
(schema, "user_tmp_v"): [],
}
self._resolve_views(views, materialized)
tables = {
(schema, "users"): [
fk(["parent_user_id"], ["user_id"], "users", name="user_id_fk")
],
(schema, "dingalings"): [
fk(["id_user"], ["user_id"], "users"),
fk(
["address_id"],
["address_id"],
"email_addresses",
name="zz_email_add_id_fg",
comment="di fk comment",
),
],
(schema, "email_addresses"): [
fk(["remote_user_id"], ["user_id"], "users")
],
(schema, "comment_test"): [],
(schema, "no_constraints"): [],
(schema, "local_table"): [
fk(
["remote_id"],
["id"],
"remote_table_2",
ref_schema=config.test_schema,
)
],
(schema, "remote_table"): [
fk(["local_id"], ["id"], "local_table", ref_schema=None)
],
(schema, "remote_table_2"): [],
(schema, "noncol_idx_test_nopk"): [],
(schema, "noncol_idx_test_pk"): [],
(schema, self.temp_table_name()): [],
}
if not testing.requires.self_referential_foreign_keys.enabled:
tables[(schema, "users")].clear()
if not testing.requires.named_constraints.enabled:
for vals in tables.values():
for val in vals:
if val["name"] is not mock.ANY:
val["name"] = mock.ANY
res = self._resolve_kind(kind, tables, views, materialized)
res = self._resolve_names(schema, scope, filter_names, res)
return res
@property
def _required_fk_keys(self):
return {
"name",
"constrained_columns",
"referred_schema",
"referred_table",
"referred_columns",
}
def exp_indexes(
self,
schema=None,
scope=ObjectScope.ANY,
kind=ObjectKind.ANY,
filter_names=None,
):
def idx(
*cols,
name,
unique=False,
column_sorting=None,
duplicates=False,
fk=False,
):
fk_req = testing.requires.foreign_keys_reflect_as_index
dup_req = testing.requires.unique_constraints_reflect_as_index
sorting_expression = (
testing.requires.reflect_indexes_with_ascdesc_as_expression
)
if (fk and not fk_req.enabled) or (
duplicates and not dup_req.enabled
):
return ()
res = {
"unique": unique,
"column_names": list(cols),
"name": name,
"dialect_options": mock.ANY,
"include_columns": [],
}
if column_sorting:
res["column_sorting"] = column_sorting
if sorting_expression.enabled:
res["expressions"] = orig = res["column_names"]
res["column_names"] = [
None if c in column_sorting else c for c in orig
]
if duplicates:
res["duplicates_constraint"] = name
return [res]
materialized = {(schema, "dingalings_v"): []}
views = {
(schema, "email_addresses_v"): [],
(schema, "users_v"): [],
(schema, "user_tmp_v"): [],
}
self._resolve_views(views, materialized)
if materialized:
materialized[(schema, "dingalings_v")].extend(
idx("data", name="mat_index")
)
tables = {
(schema, "users"): [
*idx("parent_user_id", name="user_id_fk", fk=True),
*idx("user_id", "test2", "test1", name="users_all_idx"),
*idx("test1", "test2", name="users_t_idx", unique=True),
],
(schema, "dingalings"): [
*idx("data", name=mock.ANY, unique=True, duplicates=True),
*idx("id_user", name=mock.ANY, fk=True),
*idx(
"address_id",
"dingaling_id",
name="zz_dingalings_multiple",
unique=True,
duplicates=True,
),
],
(schema, "email_addresses"): [
*idx("email_address", name=mock.ANY),
*idx("remote_user_id", name=mock.ANY, fk=True),
],
(schema, "comment_test"): [],
(schema, "no_constraints"): [],
(schema, "local_table"): [
*idx("remote_id", name=mock.ANY, fk=True)
],
(schema, "remote_table"): [
*idx("local_id", name=mock.ANY, fk=True)
],
(schema, "remote_table_2"): [],
(schema, "noncol_idx_test_nopk"): [
*idx(
"q",
name="noncol_idx_nopk",
column_sorting={"q": ("desc",)},
)
],
(schema, "noncol_idx_test_pk"): [
*idx(
"q", name="noncol_idx_pk", column_sorting={"q": ("desc",)}
)
],
(schema, self.temp_table_name()): [
*idx("foo", name="user_tmp_ix"),
*idx(
"name",
name=f"user_tmp_uq_{config.ident}",
duplicates=True,
unique=True,
),
],
}
if (
not testing.requires.indexes_with_ascdesc.enabled
or not testing.requires.reflect_indexes_with_ascdesc.enabled
):
tables[(schema, "noncol_idx_test_nopk")].clear()
tables[(schema, "noncol_idx_test_pk")].clear()
res = self._resolve_kind(kind, tables, views, materialized)
res = self._resolve_names(schema, scope, filter_names, res)
return res
@property
def _required_index_keys(self):
return {"name", "column_names", "unique"}
def exp_ucs(
self,
schema=None,
scope=ObjectScope.ANY,
kind=ObjectKind.ANY,
filter_names=None,
all_=False,
):
def uc(
*cols, name, duplicates_index=None, is_index=False, comment=None
):
req = testing.requires.unique_index_reflect_as_unique_constraints
if is_index and not req.enabled:
return ()
res = {
"column_names": list(cols),
"name": name,
"comment": comment,
}
if duplicates_index:
res["duplicates_index"] = duplicates_index
return [res]
materialized = {(schema, "dingalings_v"): []}
views = {
(schema, "email_addresses_v"): [],
(schema, "users_v"): [],
(schema, "user_tmp_v"): [],
}
self._resolve_views(views, materialized)
tables = {
(schema, "users"): [
*uc(
"test1",
"test2",
name="users_t_idx",
duplicates_index="users_t_idx",
is_index=True,
)
],
(schema, "dingalings"): [
*uc("data", name=mock.ANY, duplicates_index=mock.ANY),
*uc(
"address_id",
"dingaling_id",
name="zz_dingalings_multiple",
duplicates_index="zz_dingalings_multiple",
comment="di unique comment",
),
],
(schema, "email_addresses"): [],
(schema, "comment_test"): [],
(schema, "no_constraints"): [],
(schema, "local_table"): [],
(schema, "remote_table"): [],
(schema, "remote_table_2"): [],
(schema, "noncol_idx_test_nopk"): [],
(schema, "noncol_idx_test_pk"): [],
(schema, self.temp_table_name()): [
*uc("name", name=f"user_tmp_uq_{config.ident}")
],
}
if all_:
return {**materialized, **views, **tables}
else:
res = self._resolve_kind(kind, tables, views, materialized)
res = self._resolve_names(schema, scope, filter_names, res)
return res
@property
def _required_unique_cst_keys(self):
return {"name", "column_names"}
def exp_ccs(
self,
schema=None,
scope=ObjectScope.ANY,
kind=ObjectKind.ANY,
filter_names=None,
):
class tt(str):
def __eq__(self, other):
res = (
other.lower()
.replace("(", "")
.replace(")", "")
.replace("`", "")
)
return self in res
def cc(text, name, comment=None):
return {"sqltext": tt(text), "name": name, "comment": comment}
# print({1: "test2 > (0)::double precision"} == {1: tt("test2 > 0")})
# assert 0
materialized = {(schema, "dingalings_v"): []}
views = {
(schema, "email_addresses_v"): [],
(schema, "users_v"): [],
(schema, "user_tmp_v"): [],
}
self._resolve_views(views, materialized)
tables = {
(schema, "users"): [
cc("test2 <= 1000", mock.ANY),
cc(
"test2 > 0",
"zz_test2_gt_zero",
comment="users check constraint",
),
],
(schema, "dingalings"): [
cc(
"address_id > 0 and address_id < 1000",
name="address_id_gt_zero",
),
],
(schema, "email_addresses"): [],
(schema, "comment_test"): [],
(schema, "no_constraints"): [],
(schema, "local_table"): [],
(schema, "remote_table"): [],
(schema, "remote_table_2"): [],
(schema, "noncol_idx_test_nopk"): [],
(schema, "noncol_idx_test_pk"): [],
(schema, self.temp_table_name()): [],
}
res = self._resolve_kind(kind, tables, views, materialized)
res = self._resolve_names(schema, scope, filter_names, res)
return res
@property
def _required_cc_keys(self):
return {"name", "sqltext"}
@testing.requires.schema_reflection
def test_get_schema_names(self, connection):
insp = inspect(connection)
is_true(testing.config.test_schema in insp.get_schema_names())
@testing.requires.schema_reflection
def test_has_schema(self, connection):
insp = inspect(connection)
is_true(insp.has_schema(testing.config.test_schema))
is_false(insp.has_schema("sa_fake_schema_foo"))
@testing.requires.schema_reflection
def test_get_schema_names_w_translate_map(self, connection):
"""test #7300"""
connection = connection.execution_options(
schema_translate_map={
"foo": "bar",
BLANK_SCHEMA: testing.config.test_schema,
}
)
insp = inspect(connection)
is_true(testing.config.test_schema in insp.get_schema_names())
@testing.requires.schema_reflection
def test_has_schema_w_translate_map(self, connection):
connection = connection.execution_options(
schema_translate_map={
"foo": "bar",
BLANK_SCHEMA: testing.config.test_schema,
}
)
insp = inspect(connection)
is_true(insp.has_schema(testing.config.test_schema))
is_false(insp.has_schema("sa_fake_schema_foo"))
@testing.requires.schema_reflection
@testing.requires.schema_create_delete
def test_schema_cache(self, connection):
insp = inspect(connection)
is_false("foo_bar" in insp.get_schema_names())
is_false(insp.has_schema("foo_bar"))
connection.execute(DDL("CREATE SCHEMA foo_bar"))
try:
is_false("foo_bar" in insp.get_schema_names())
is_false(insp.has_schema("foo_bar"))
insp.clear_cache()
is_true("foo_bar" in insp.get_schema_names())
is_true(insp.has_schema("foo_bar"))
finally:
connection.execute(DDL("DROP SCHEMA foo_bar"))
@testing.requires.schema_reflection
def test_dialect_initialize(self):
engine = engines.testing_engine()
inspect(engine)
assert hasattr(engine.dialect, "default_schema_name")
@testing.requires.schema_reflection
def test_get_default_schema_name(self, connection):
insp = inspect(connection)
eq_(insp.default_schema_name, connection.dialect.default_schema_name)
@testing.combinations(
None,
("foreign_key", testing.requires.foreign_key_constraint_reflection),
argnames="order_by",
)
@testing.combinations(
(True, testing.requires.schemas), False, argnames="use_schema"
)
def test_get_table_names(self, connection, order_by, use_schema):
if use_schema:
schema = config.test_schema
else:
schema = None
_ignore_tables = {
"comment_test",
"noncol_idx_test_pk",
"noncol_idx_test_nopk",
"local_table",
"remote_table",
"remote_table_2",
"no_constraints",
}
insp = inspect(connection)
if order_by:
tables = [
rec[0]
for rec in insp.get_sorted_table_and_fkc_names(schema)
if rec[0]
]
else:
tables = insp.get_table_names(schema)
table_names = [t for t in tables if t not in _ignore_tables]
if order_by == "foreign_key":
answer = ["users", "email_addresses", "dingalings"]
eq_(table_names, answer)
else:
answer = ["dingalings", "email_addresses", "users"]
eq_(sorted(table_names), answer)
@testing.combinations(
(True, testing.requires.schemas), False, argnames="use_schema"
)
def test_get_view_names(self, connection, use_schema):
insp = inspect(connection)
if use_schema:
schema = config.test_schema
else:
schema = None
table_names = insp.get_view_names(schema)
if testing.requires.materialized_views.enabled:
eq_(sorted(table_names), ["email_addresses_v", "users_v"])
eq_(insp.get_materialized_view_names(schema), ["dingalings_v"])
else:
answer = ["dingalings_v", "email_addresses_v", "users_v"]
eq_(sorted(table_names), answer)
@testing.requires.temp_table_names
def test_get_temp_table_names(self, connection):
insp = inspect(connection)
temp_table_names = insp.get_temp_table_names()
eq_(sorted(temp_table_names), [f"user_tmp_{config.ident}"])
@testing.requires.view_reflection
@testing.requires.temporary_views
def test_get_temp_view_names(self, connection):
insp = inspect(connection)
temp_table_names = insp.get_temp_view_names()
eq_(sorted(temp_table_names), ["user_tmp_v"])
@testing.requires.comment_reflection
def test_get_comments(self, connection):
self._test_get_comments(connection)
@testing.requires.comment_reflection
@testing.requires.schemas
def test_get_comments_with_schema(self, connection):
self._test_get_comments(connection, testing.config.test_schema)
def _test_get_comments(self, connection, schema=None):
insp = inspect(connection)
exp = self.exp_comments(schema=schema)
eq_(
insp.get_table_comment("comment_test", schema=schema),
exp[(schema, "comment_test")],
)
eq_(
insp.get_table_comment("users", schema=schema),
exp[(schema, "users")],
)
eq_(
insp.get_table_comment("comment_test", schema=schema),
exp[(schema, "comment_test")],
)
no_cst = self.tables.no_constraints.name
eq_(
insp.get_table_comment(no_cst, schema=schema),
exp[(schema, no_cst)],
)
@testing.combinations(
(False, False),
(False, True, testing.requires.schemas),
(True, False, testing.requires.view_reflection),
(
True,
True,
testing.requires.schemas + testing.requires.view_reflection,
),
argnames="use_views,use_schema",
)
def test_get_columns(self, connection, use_views, use_schema):
if use_schema:
schema = config.test_schema
else:
schema = None
users, addresses = (self.tables.users, self.tables.email_addresses)
if use_views:
table_names = ["users_v", "email_addresses_v", "dingalings_v"]
else:
table_names = ["users", "email_addresses"]
insp = inspect(connection)
for table_name, table in zip(table_names, (users, addresses)):
schema_name = schema
cols = insp.get_columns(table_name, schema=schema_name)
is_true(len(cols) > 0, len(cols))
# should be in order
for i, col in enumerate(table.columns):
eq_(col.name, cols[i]["name"])
ctype = cols[i]["type"].__class__
ctype_def = col.type
if isinstance(ctype_def, sa.types.TypeEngine):
ctype_def = ctype_def.__class__
# Oracle returns Date for DateTime.
if testing.against("oracle") and ctype_def in (
sql_types.Date,
sql_types.DateTime,
):
ctype_def = sql_types.Date
# assert that the desired type and return type share
# a base within one of the generic types.
is_true(
len(
set(ctype.__mro__)
.intersection(ctype_def.__mro__)
.intersection(
[
sql_types.Integer,
sql_types.Numeric,
sql_types.DateTime,
sql_types.Date,
sql_types.Time,
sql_types.String,
sql_types._Binary,
]
)
)
> 0,
"%s(%s), %s(%s)"
% (col.name, col.type, cols[i]["name"], ctype),
)
if not col.primary_key:
assert cols[i]["default"] is None
# The case of a table with no column
# is tested below in TableNoColumnsTest
@testing.requires.temp_table_reflection
def test_reflect_table_temp_table(self, connection):
table_name = self.temp_table_name()
user_tmp = self.tables[table_name]
reflected_user_tmp = Table(
table_name, MetaData(), autoload_with=connection
)
self.assert_tables_equal(
user_tmp, reflected_user_tmp, strict_constraints=False
)
@testing.requires.temp_table_reflection
def test_get_temp_table_columns(self, connection):
table_name = self.temp_table_name()
user_tmp = self.tables[table_name]
insp = inspect(connection)
cols = insp.get_columns(table_name)
is_true(len(cols) > 0, len(cols))
for i, col in enumerate(user_tmp.columns):
eq_(col.name, cols[i]["name"])
@testing.requires.temp_table_reflection
@testing.requires.view_column_reflection
@testing.requires.temporary_views
def test_get_temp_view_columns(self, connection):
insp = inspect(connection)
cols = insp.get_columns("user_tmp_v")
eq_([col["name"] for col in cols], ["id", "name", "foo"])
@testing.combinations(
(False,), (True, testing.requires.schemas), argnames="use_schema"
)
@testing.requires.primary_key_constraint_reflection
def test_get_pk_constraint(self, connection, use_schema):
if use_schema:
schema = testing.config.test_schema
else:
schema = None
users, addresses = self.tables.users, self.tables.email_addresses
insp = inspect(connection)
exp = self.exp_pks(schema=schema)
users_cons = insp.get_pk_constraint(users.name, schema=schema)
self._check_list(
[users_cons], [exp[(schema, users.name)]], self._required_pk_keys
)
addr_cons = insp.get_pk_constraint(addresses.name, schema=schema)
exp_cols = exp[(schema, addresses.name)]["constrained_columns"]
eq_(addr_cons["constrained_columns"], exp_cols)
with testing.requires.reflects_pk_names.fail_if():
eq_(addr_cons["name"], "email_ad_pk")
no_cst = self.tables.no_constraints.name
self._check_list(
[insp.get_pk_constraint(no_cst, schema=schema)],
[exp[(schema, no_cst)]],
self._required_pk_keys,
)
@testing.combinations(
(False,), (True, testing.requires.schemas), argnames="use_schema"
)
@testing.requires.foreign_key_constraint_reflection
def test_get_foreign_keys(self, connection, use_schema):
if use_schema:
schema = config.test_schema
else:
schema = None
users, addresses = (self.tables.users, self.tables.email_addresses)
insp = inspect(connection)
expected_schema = schema
# users
if testing.requires.self_referential_foreign_keys.enabled:
users_fkeys = insp.get_foreign_keys(users.name, schema=schema)
fkey1 = users_fkeys[0]
with testing.requires.named_constraints.fail_if():
eq_(fkey1["name"], "user_id_fk")
eq_(fkey1["referred_schema"], expected_schema)
eq_(fkey1["referred_table"], users.name)
eq_(fkey1["referred_columns"], ["user_id"])
eq_(fkey1["constrained_columns"], ["parent_user_id"])
# addresses
addr_fkeys = insp.get_foreign_keys(addresses.name, schema=schema)
fkey1 = addr_fkeys[0]
with testing.requires.implicitly_named_constraints.fail_if():
is_true(fkey1["name"] is not None)
eq_(fkey1["referred_schema"], expected_schema)
eq_(fkey1["referred_table"], users.name)
eq_(fkey1["referred_columns"], ["user_id"])
eq_(fkey1["constrained_columns"], ["remote_user_id"])
no_cst = self.tables.no_constraints.name
eq_(insp.get_foreign_keys(no_cst, schema=schema), [])
@testing.requires.cross_schema_fk_reflection
@testing.requires.schemas
def test_get_inter_schema_foreign_keys(self, connection):
local_table, remote_table, remote_table_2 = self.tables(
"%s.local_table" % connection.dialect.default_schema_name,
"%s.remote_table" % testing.config.test_schema,
"%s.remote_table_2" % testing.config.test_schema,
)
insp = inspect(connection)
local_fkeys = insp.get_foreign_keys(local_table.name)
eq_(len(local_fkeys), 1)
fkey1 = local_fkeys[0]
eq_(fkey1["referred_schema"], testing.config.test_schema)
eq_(fkey1["referred_table"], remote_table_2.name)
eq_(fkey1["referred_columns"], ["id"])
eq_(fkey1["constrained_columns"], ["remote_id"])
remote_fkeys = insp.get_foreign_keys(
remote_table.name, schema=testing.config.test_schema
)
eq_(len(remote_fkeys), 1)
fkey2 = remote_fkeys[0]
is_true(
fkey2["referred_schema"]
in (
None,
connection.dialect.default_schema_name,
)
)
eq_(fkey2["referred_table"], local_table.name)
eq_(fkey2["referred_columns"], ["id"])
eq_(fkey2["constrained_columns"], ["local_id"])
@testing.combinations(
(False,), (True, testing.requires.schemas), argnames="use_schema"
)
@testing.requires.index_reflection
def test_get_indexes(self, connection, use_schema):
if use_schema:
schema = config.test_schema
else:
schema = None
# The database may decide to create indexes for foreign keys, etc.
# so there may be more indexes than expected.
insp = inspect(connection)
indexes = insp.get_indexes("users", schema=schema)
exp = self.exp_indexes(schema=schema)
self._check_list(
indexes, exp[(schema, "users")], self._required_index_keys
)
no_cst = self.tables.no_constraints.name
self._check_list(
insp.get_indexes(no_cst, schema=schema),
exp[(schema, no_cst)],
self._required_index_keys,
)
@testing.combinations(
("noncol_idx_test_nopk", "noncol_idx_nopk"),
("noncol_idx_test_pk", "noncol_idx_pk"),
argnames="tname,ixname",
)
@testing.requires.index_reflection
@testing.requires.indexes_with_ascdesc
@testing.requires.reflect_indexes_with_ascdesc
def test_get_noncol_index(self, connection, tname, ixname):
insp = inspect(connection)
indexes = insp.get_indexes(tname)
# reflecting an index that has "x DESC" in it as the column.
# the DB may or may not give us "x", but make sure we get the index
# back, it has a name, it's connected to the table.
expected_indexes = self.exp_indexes()[(None, tname)]
self._check_list(indexes, expected_indexes, self._required_index_keys)
t = Table(tname, MetaData(), autoload_with=connection)
eq_(len(t.indexes), 1)
is_(list(t.indexes)[0].table, t)
eq_(list(t.indexes)[0].name, ixname)
@testing.requires.temp_table_reflection
@testing.requires.unique_constraint_reflection
def test_get_temp_table_unique_constraints(self, connection):
insp = inspect(connection)
name = self.temp_table_name()
reflected = insp.get_unique_constraints(name)
exp = self.exp_ucs(all_=True)[(None, name)]
self._check_list(reflected, exp, self._required_index_keys)
@testing.requires.temp_table_reflect_indexes
def test_get_temp_table_indexes(self, connection):
insp = inspect(connection)
table_name = self.temp_table_name()
indexes = insp.get_indexes(table_name)
for ind in indexes:
ind.pop("dialect_options", None)
expected = [
{"unique": False, "column_names": ["foo"], "name": "user_tmp_ix"}
]
if testing.requires.index_reflects_included_columns.enabled:
expected[0]["include_columns"] = []
eq_(
[idx for idx in indexes if idx["name"] == "user_tmp_ix"],
expected,
)
@testing.combinations(
(True, testing.requires.schemas), (False,), argnames="use_schema"
)
@testing.requires.unique_constraint_reflection
def test_get_unique_constraints(self, metadata, connection, use_schema):
# SQLite dialect needs to parse the names of the constraints
# separately from what it gets from PRAGMA index_list(), and
# then matches them up. so same set of column_names in two
# constraints will confuse it. Perhaps we should no longer
# bother with index_list() here since we have the whole
# CREATE TABLE?
if use_schema:
schema = config.test_schema
else:
schema = None
uniques = sorted(
[
{"name": "unique_a", "column_names": ["a"]},
{"name": "unique_a_b_c", "column_names": ["a", "b", "c"]},
{"name": "unique_c_a_b", "column_names": ["c", "a", "b"]},
{"name": "unique_asc_key", "column_names": ["asc", "key"]},
{"name": "i.have.dots", "column_names": ["b"]},
{"name": "i have spaces", "column_names": ["c"]},
],
key=operator.itemgetter("name"),
)
table = Table(
"testtbl",
metadata,
Column("a", sa.String(20)),
Column("b", sa.String(30)),
Column("c", sa.Integer),
# reserved identifiers
Column("asc", sa.String(30)),
Column("key", sa.String(30)),
schema=schema,
)
for uc in uniques:
table.append_constraint(
sa.UniqueConstraint(*uc["column_names"], name=uc["name"])
)
table.create(connection)
insp = inspect(connection)
reflected = sorted(
insp.get_unique_constraints("testtbl", schema=schema),
key=operator.itemgetter("name"),
)
names_that_duplicate_index = set()
eq_(len(uniques), len(reflected))
for orig, refl in zip(uniques, reflected):
# Different dialects handle duplicate index and constraints
# differently, so ignore this flag
dupe = refl.pop("duplicates_index", None)
if dupe:
names_that_duplicate_index.add(dupe)
eq_(refl.pop("comment", None), None)
eq_(orig, refl)
reflected_metadata = MetaData()
reflected = Table(
"testtbl",
reflected_metadata,
autoload_with=connection,
schema=schema,
)
# test "deduplicates for index" logic. MySQL and Oracle
# "unique constraints" are actually unique indexes (with possible
# exception of a unique that is a dupe of another one in the case
# of Oracle). make sure # they aren't duplicated.
idx_names = {idx.name for idx in reflected.indexes}
uq_names = {
uq.name
for uq in reflected.constraints
if isinstance(uq, sa.UniqueConstraint)
}.difference(["unique_c_a_b"])
assert not idx_names.intersection(uq_names)
if names_that_duplicate_index:
eq_(names_that_duplicate_index, idx_names)
eq_(uq_names, set())
no_cst = self.tables.no_constraints.name
eq_(insp.get_unique_constraints(no_cst, schema=schema), [])
@testing.requires.view_reflection
@testing.combinations(
(False,), (True, testing.requires.schemas), argnames="use_schema"
)
def test_get_view_definition(self, connection, use_schema):
if use_schema:
schema = config.test_schema
else:
schema = None
insp = inspect(connection)
for view in ["users_v", "email_addresses_v", "dingalings_v"]:
v = insp.get_view_definition(view, schema=schema)
is_true(bool(v))
@testing.requires.view_reflection
def test_get_view_definition_does_not_exist(self, connection):
insp = inspect(connection)
with expect_raises(NoSuchTableError):
insp.get_view_definition("view_does_not_exist")
with expect_raises(NoSuchTableError):
insp.get_view_definition("users") # a table
@testing.requires.table_reflection
def test_autoincrement_col(self, connection):
"""test that 'autoincrement' is reflected according to sqla's policy.
Don't mark this test as unsupported for any backend !
(technically it fails with MySQL InnoDB since "id" comes before "id2")
A backend is better off not returning "autoincrement" at all,
instead of potentially returning "False" for an auto-incrementing
primary key column.
"""
insp = inspect(connection)
for tname, cname in [
("users", "user_id"),
("email_addresses", "address_id"),
("dingalings", "dingaling_id"),
]:
cols = insp.get_columns(tname)
id_ = {c["name"]: c for c in cols}[cname]
assert id_.get("autoincrement", True)
@testing.combinations(
(True, testing.requires.schemas), (False,), argnames="use_schema"
)
def test_get_table_options(self, use_schema):
insp = inspect(config.db)
schema = config.test_schema if use_schema else None
if testing.requires.reflect_table_options.enabled:
res = insp.get_table_options("users", schema=schema)
is_true(isinstance(res, dict))
# NOTE: can't really create a table with no option
res = insp.get_table_options("no_constraints", schema=schema)
is_true(isinstance(res, dict))
else:
with expect_raises(NotImplementedError):
res = insp.get_table_options("users", schema=schema)
@testing.combinations((True, testing.requires.schemas), False)
def test_multi_get_table_options(self, use_schema):
insp = inspect(config.db)
if testing.requires.reflect_table_options.enabled:
schema = config.test_schema if use_schema else None
res = insp.get_multi_table_options(schema=schema)
exp = {
(schema, table): insp.get_table_options(table, schema=schema)
for table in insp.get_table_names(schema=schema)
}
eq_(res, exp)
else:
with expect_raises(NotImplementedError):
res = insp.get_multi_table_options()
@testing.fixture
def get_multi_exp(self, connection):
def provide_fixture(
schema, scope, kind, use_filter, single_reflect_fn, exp_method
):
insp = inspect(connection)
# call the reflection function at least once to avoid
# "Unexpected success" errors if the result is actually empty
# and NotImplementedError is not raised
single_reflect_fn(insp, "email_addresses")
kw = {"scope": scope, "kind": kind}
if schema:
schema = schema()
filter_names = []
if ObjectKind.TABLE in kind:
filter_names.extend(
["comment_test", "users", "does-not-exist"]
)
if ObjectKind.VIEW in kind:
filter_names.extend(["email_addresses_v", "does-not-exist"])
if ObjectKind.MATERIALIZED_VIEW in kind:
filter_names.extend(["dingalings_v", "does-not-exist"])
if schema:
kw["schema"] = schema
if use_filter:
kw["filter_names"] = filter_names
exp = exp_method(
schema=schema,
scope=scope,
kind=kind,
filter_names=kw.get("filter_names"),
)
kws = [kw]
if scope == ObjectScope.DEFAULT:
nkw = kw.copy()
nkw.pop("scope")
kws.append(nkw)
if kind == ObjectKind.TABLE:
nkw = kw.copy()
nkw.pop("kind")
kws.append(nkw)
return inspect(connection), kws, exp
return provide_fixture
@testing.requires.reflect_table_options
@_multi_combination
def test_multi_get_table_options_tables(
self, get_multi_exp, schema, scope, kind, use_filter
):
insp, kws, exp = get_multi_exp(
schema,
scope,
kind,
use_filter,
Inspector.get_table_options,
self.exp_options,
)
for kw in kws:
insp.clear_cache()
result = insp.get_multi_table_options(**kw)
eq_(result, exp)
@testing.requires.comment_reflection
@_multi_combination
def test_get_multi_table_comment(
self, get_multi_exp, schema, scope, kind, use_filter
):
insp, kws, exp = get_multi_exp(
schema,
scope,
kind,
use_filter,
Inspector.get_table_comment,
self.exp_comments,
)
for kw in kws:
insp.clear_cache()
eq_(insp.get_multi_table_comment(**kw), exp)
def _check_expressions(self, result, exp, err_msg):
def _clean(text: str):
return re.sub(r"['\" ]", "", text).lower()
if isinstance(exp, dict):
eq_({_clean(e): v for e, v in result.items()}, exp, err_msg)
else:
eq_([_clean(e) for e in result], exp, err_msg)
def _check_list(self, result, exp, req_keys=None, msg=None):
if req_keys is None:
eq_(result, exp, msg)
else:
eq_(len(result), len(exp), msg)
for r, e in zip(result, exp):
for k in set(r) | set(e):
if k in req_keys or (k in r and k in e):
err_msg = f"{msg} - {k} - {r}"
if k in ("expressions", "column_sorting"):
self._check_expressions(r[k], e[k], err_msg)
else:
eq_(r[k], e[k], err_msg)
def _check_table_dict(self, result, exp, req_keys=None, make_lists=False):
eq_(set(result.keys()), set(exp.keys()))
for k in result:
r, e = result[k], exp[k]
if make_lists:
r, e = [r], [e]
self._check_list(r, e, req_keys, k)
@_multi_combination
def test_get_multi_columns(
self, get_multi_exp, schema, scope, kind, use_filter
):
insp, kws, exp = get_multi_exp(
schema,
scope,
kind,
use_filter,
Inspector.get_columns,
self.exp_columns,
)
for kw in kws:
insp.clear_cache()
result = insp.get_multi_columns(**kw)
self._check_table_dict(result, exp, self._required_column_keys)
@testing.requires.primary_key_constraint_reflection
@_multi_combination
def test_get_multi_pk_constraint(
self, get_multi_exp, schema, scope, kind, use_filter
):
insp, kws, exp = get_multi_exp(
schema,
scope,
kind,
use_filter,
Inspector.get_pk_constraint,
self.exp_pks,
)
for kw in kws:
insp.clear_cache()
result = insp.get_multi_pk_constraint(**kw)
self._check_table_dict(
result, exp, self._required_pk_keys, make_lists=True
)
def _adjust_sort(self, result, expected, key):
if not testing.requires.implicitly_named_constraints.enabled:
for obj in [result, expected]:
for val in obj.values():
if len(val) > 1 and any(
v.get("name") in (None, mock.ANY) for v in val
):
val.sort(key=key)
@testing.requires.foreign_key_constraint_reflection
@_multi_combination
def test_get_multi_foreign_keys(
self, get_multi_exp, schema, scope, kind, use_filter
):
insp, kws, exp = get_multi_exp(
schema,
scope,
kind,
use_filter,
Inspector.get_foreign_keys,
self.exp_fks,
)
for kw in kws:
insp.clear_cache()
result = insp.get_multi_foreign_keys(**kw)
self._adjust_sort(
result, exp, lambda d: tuple(d["constrained_columns"])
)
self._check_table_dict(result, exp, self._required_fk_keys)
@testing.requires.index_reflection
@_multi_combination
def test_get_multi_indexes(
self, get_multi_exp, schema, scope, kind, use_filter
):
insp, kws, exp = get_multi_exp(
schema,
scope,
kind,
use_filter,
Inspector.get_indexes,
self.exp_indexes,
)
for kw in kws:
insp.clear_cache()
result = insp.get_multi_indexes(**kw)
self._check_table_dict(result, exp, self._required_index_keys)
@testing.requires.unique_constraint_reflection
@_multi_combination
def test_get_multi_unique_constraints(
self, get_multi_exp, schema, scope, kind, use_filter
):
insp, kws, exp = get_multi_exp(
schema,
scope,
kind,
use_filter,
Inspector.get_unique_constraints,
self.exp_ucs,
)
for kw in kws:
insp.clear_cache()
result = insp.get_multi_unique_constraints(**kw)
self._adjust_sort(result, exp, lambda d: tuple(d["column_names"]))
self._check_table_dict(result, exp, self._required_unique_cst_keys)
@testing.requires.check_constraint_reflection
@_multi_combination
def test_get_multi_check_constraints(
self, get_multi_exp, schema, scope, kind, use_filter
):
insp, kws, exp = get_multi_exp(
schema,
scope,
kind,
use_filter,
Inspector.get_check_constraints,
self.exp_ccs,
)
for kw in kws:
insp.clear_cache()
result = insp.get_multi_check_constraints(**kw)
self._adjust_sort(result, exp, lambda d: tuple(d["sqltext"]))
self._check_table_dict(result, exp, self._required_cc_keys)
@testing.combinations(
("get_table_options", testing.requires.reflect_table_options),
"get_columns",
(
"get_pk_constraint",
testing.requires.primary_key_constraint_reflection,
),
(
"get_foreign_keys",
testing.requires.foreign_key_constraint_reflection,
),
("get_indexes", testing.requires.index_reflection),
(
"get_unique_constraints",
testing.requires.unique_constraint_reflection,
),
(
"get_check_constraints",
testing.requires.check_constraint_reflection,
),
("get_table_comment", testing.requires.comment_reflection),
argnames="method",
)
def test_not_existing_table(self, method, connection):
insp = inspect(connection)
meth = getattr(insp, method)
with expect_raises(NoSuchTableError):
meth("table_does_not_exists")
def test_unreflectable(self, connection):
mc = Inspector.get_multi_columns
def patched(*a, **k):
ur = k.setdefault("unreflectable", {})
ur[(None, "some_table")] = UnreflectableTableError("err")
return mc(*a, **k)
with mock.patch.object(Inspector, "get_multi_columns", patched):
with expect_raises_message(UnreflectableTableError, "err"):
inspect(connection).reflect_table(
Table("some_table", MetaData()), None
)
@testing.combinations(True, False, argnames="use_schema")
@testing.combinations(
(True, testing.requires.views), False, argnames="views"
)
def test_metadata(self, connection, use_schema, views):
m = MetaData()
schema = config.test_schema if use_schema else None
m.reflect(connection, schema=schema, views=views, resolve_fks=False)
insp = inspect(connection)
tables = insp.get_table_names(schema)
if views:
tables += insp.get_view_names(schema)
try:
tables += insp.get_materialized_view_names(schema)
except NotImplementedError:
pass
if schema:
tables = [f"{schema}.{t}" for t in tables]
eq_(sorted(m.tables), sorted(tables))
@testing.requires.comment_reflection
def test_comments_unicode(self, connection, metadata):
Table(
"unicode_comments",
metadata,
Column("unicode", Integer, comment="é試蛇ẟΩ"),
Column("emoji", Integer, comment="☁️✨"),
comment="試蛇ẟΩ✨",
)
metadata.create_all(connection)
insp = inspect(connection)
tc = insp.get_table_comment("unicode_comments")
eq_(tc, {"text": "試蛇ẟΩ✨"})
cols = insp.get_columns("unicode_comments")
value = {c["name"]: c["comment"] for c in cols}
exp = {"unicode": "é試蛇ẟΩ", "emoji": "☁️✨"}
eq_(value, exp)
@testing.requires.comment_reflection_full_unicode
def test_comments_unicode_full(self, connection, metadata):
Table(
"unicode_comments",
metadata,
Column("emoji", Integer, comment="🐍🧙🝝🧙♂️🧙♀️"),
comment="🎩🁰🝑🤷♀️🤷♂️",
)
metadata.create_all(connection)
insp = inspect(connection)
tc = insp.get_table_comment("unicode_comments")
eq_(tc, {"text": "🎩🁰🝑🤷♀️🤷♂️"})
c = insp.get_columns("unicode_comments")[0]
eq_({c["name"]: c["comment"]}, {"emoji": "🐍🧙🝝🧙♂️🧙♀️"})
class TableNoColumnsTest(fixtures.TestBase):
__requires__ = ("reflect_tables_no_columns",)
__backend__ = True
@testing.fixture
def table_no_columns(self, connection, metadata):
Table("empty", metadata)
metadata.create_all(connection)
@testing.fixture
def view_no_columns(self, connection, metadata):
Table("empty", metadata)
event.listen(
metadata,
"after_create",
DDL("CREATE VIEW empty_v AS SELECT * FROM empty"),
)
# for transactional DDL the transaction is rolled back before this
# drop statement is invoked
event.listen(
metadata, "before_drop", DDL("DROP VIEW IF EXISTS empty_v")
)
metadata.create_all(connection)
def test_reflect_table_no_columns(self, connection, table_no_columns):
t2 = Table("empty", MetaData(), autoload_with=connection)
eq_(list(t2.c), [])
def test_get_columns_table_no_columns(self, connection, table_no_columns):
insp = inspect(connection)
eq_(insp.get_columns("empty"), [])
multi = insp.get_multi_columns()
eq_(multi, {(None, "empty"): []})
def test_reflect_incl_table_no_columns(self, connection, table_no_columns):
m = MetaData()
m.reflect(connection)
assert set(m.tables).intersection(["empty"])
@testing.requires.views
def test_reflect_view_no_columns(self, connection, view_no_columns):
t2 = Table("empty_v", MetaData(), autoload_with=connection)
eq_(list(t2.c), [])
@testing.requires.views
def test_get_columns_view_no_columns(self, connection, view_no_columns):
insp = inspect(connection)
eq_(insp.get_columns("empty_v"), [])
multi = insp.get_multi_columns(kind=ObjectKind.VIEW)
eq_(multi, {(None, "empty_v"): []})
class ComponentReflectionTestExtra(ComparesIndexes, fixtures.TestBase):
__backend__ = True
@testing.combinations(
(True, testing.requires.schemas), (False,), argnames="use_schema"
)
@testing.requires.check_constraint_reflection
def test_get_check_constraints(self, metadata, connection, use_schema):
if use_schema:
schema = config.test_schema
else:
schema = None
Table(
"sa_cc",
metadata,
Column("a", Integer()),
sa.CheckConstraint("a > 1 AND a < 5", name="cc1"),
sa.CheckConstraint(
"a = 1 OR (a > 2 AND a < 5)", name="UsesCasing"
),
schema=schema,
)
Table(
"no_constraints",
metadata,
Column("data", sa.String(20)),
schema=schema,
)
metadata.create_all(connection)
insp = inspect(connection)
reflected = sorted(
insp.get_check_constraints("sa_cc", schema=schema),
key=operator.itemgetter("name"),
)
# trying to minimize effect of quoting, parenthesis, etc.
# may need to add more to this as new dialects get CHECK
# constraint reflection support
def normalize(sqltext):
return " ".join(
re.findall(r"and|\d|=|a|or|<|>", sqltext.lower(), re.I)
)
reflected = [
{"name": item["name"], "sqltext": normalize(item["sqltext"])}
for item in reflected
]
eq_(
reflected,
[
{"name": "UsesCasing", "sqltext": "a = 1 or a > 2 and a < 5"},
{"name": "cc1", "sqltext": "a > 1 and a < 5"},
],
)
no_cst = "no_constraints"
eq_(insp.get_check_constraints(no_cst, schema=schema), [])
@testing.requires.indexes_with_expressions
def test_reflect_expression_based_indexes(self, metadata, connection):
t = Table(
"t",
metadata,
Column("x", String(30)),
Column("y", String(30)),
Column("z", String(30)),
)
Index("t_idx", func.lower(t.c.x), t.c.z, func.lower(t.c.y))
long_str = "long string " * 100
Index("t_idx_long", func.coalesce(t.c.x, long_str))
Index("t_idx_2", t.c.x)
metadata.create_all(connection)
insp = inspect(connection)
expected = [
{
"name": "t_idx_2",
"column_names": ["x"],
"unique": False,
"dialect_options": {},
}
]
def completeIndex(entry):
if testing.requires.index_reflects_included_columns.enabled:
entry["include_columns"] = []
entry["dialect_options"] = {
f"{connection.engine.name}_include": []
}
else:
entry.setdefault("dialect_options", {})
completeIndex(expected[0])
class lower_index_str(str):
def __eq__(self, other):
ol = other.lower()
# test that lower and x or y are in the string
return "lower" in ol and ("x" in ol or "y" in ol)
class coalesce_index_str(str):
def __eq__(self, other):
# test that coalesce and the string is in other
return "coalesce" in other.lower() and long_str in other
if testing.requires.reflect_indexes_with_expressions.enabled:
expr_index = {
"name": "t_idx",
"column_names": [None, "z", None],
"expressions": [
lower_index_str("lower(x)"),
"z",
lower_index_str("lower(y)"),
],
"unique": False,
}
completeIndex(expr_index)
expected.insert(0, expr_index)
expr_index_long = {
"name": "t_idx_long",
"column_names": [None],
"expressions": [
coalesce_index_str(f"coalesce(x, '{long_str}')")
],
"unique": False,
}
completeIndex(expr_index_long)
expected.append(expr_index_long)
eq_(insp.get_indexes("t"), expected)
m2 = MetaData()
t2 = Table("t", m2, autoload_with=connection)
else:
with expect_warnings(
"Skipped unsupported reflection of expression-based "
"index t_idx"
):
eq_(insp.get_indexes("t"), expected)
m2 = MetaData()
t2 = Table("t", m2, autoload_with=connection)
self.compare_table_index_with_expected(
t2, expected, connection.engine.name
)
@testing.requires.index_reflects_included_columns
def test_reflect_covering_index(self, metadata, connection):
t = Table(
"t",
metadata,
Column("x", String(30)),
Column("y", String(30)),
)
idx = Index("t_idx", t.c.x)
idx.dialect_options[connection.engine.name]["include"] = ["y"]
metadata.create_all(connection)
insp = inspect(connection)
get_indexes = insp.get_indexes("t")
eq_(
get_indexes,
[
{
"name": "t_idx",
"column_names": ["x"],
"include_columns": ["y"],
"unique": False,
"dialect_options": mock.ANY,
}
],
)
eq_(
get_indexes[0]["dialect_options"][
"%s_include" % connection.engine.name
],
["y"],
)
t2 = Table("t", MetaData(), autoload_with=connection)
eq_(
list(t2.indexes)[0].dialect_options[connection.engine.name][
"include"
],
["y"],
)
def _type_round_trip(self, connection, metadata, *types):
t = Table(
"t",
metadata,
*[Column("t%d" % i, type_) for i, type_ in enumerate(types)],
)
t.create(connection)
return [c["type"] for c in inspect(connection).get_columns("t")]
@testing.requires.table_reflection
def test_numeric_reflection(self, connection, metadata):
for typ in self._type_round_trip(
connection, metadata, sql_types.Numeric(18, 5)
):
assert isinstance(typ, sql_types.Numeric)
eq_(typ.precision, 18)
eq_(typ.scale, 5)
@testing.requires.table_reflection
def test_varchar_reflection(self, connection, metadata):
typ = self._type_round_trip(
connection, metadata, sql_types.String(52)
)[0]
assert isinstance(typ, sql_types.String)
eq_(typ.length, 52)
@testing.requires.table_reflection
def test_nullable_reflection(self, connection, metadata):
t = Table(
"t",
metadata,
Column("a", Integer, nullable=True),
Column("b", Integer, nullable=False),
)
t.create(connection)
eq_(
{
col["name"]: col["nullable"]
for col in inspect(connection).get_columns("t")
},
{"a": True, "b": False},
)
@testing.combinations(
(
None,
"CASCADE",
None,
testing.requires.foreign_key_constraint_option_reflection_ondelete,
),
(
None,
None,
"SET NULL",
testing.requires.foreign_key_constraint_option_reflection_onupdate,
),
(
{},
None,
"NO ACTION",
testing.requires.foreign_key_constraint_option_reflection_onupdate,
),
(
{},
"NO ACTION",
None,
testing.requires.fk_constraint_option_reflection_ondelete_noaction,
),
(
None,
None,
"RESTRICT",
testing.requires.fk_constraint_option_reflection_onupdate_restrict,
),
(
None,
"RESTRICT",
None,
testing.requires.fk_constraint_option_reflection_ondelete_restrict,
),
argnames="expected,ondelete,onupdate",
)
def test_get_foreign_key_options(
self, connection, metadata, expected, ondelete, onupdate
):
options = {}
if ondelete:
options["ondelete"] = ondelete
if onupdate:
options["onupdate"] = onupdate
if expected is None:
expected = options
Table(
"x",
metadata,
Column("id", Integer, primary_key=True),
test_needs_fk=True,
)
Table(
"table",
metadata,
Column("id", Integer, primary_key=True),
Column("x_id", Integer, ForeignKey("x.id", name="xid")),
Column("test", String(10)),
test_needs_fk=True,
)
Table(
"user",
metadata,
Column("id", Integer, primary_key=True),
Column("name", String(50), nullable=False),
Column("tid", Integer),
sa.ForeignKeyConstraint(
["tid"], ["table.id"], name="myfk", **options
),
test_needs_fk=True,
)
metadata.create_all(connection)
insp = inspect(connection)
# test 'options' is always present for a backend
# that can reflect these, since alembic looks for this
opts = insp.get_foreign_keys("table")[0]["options"]
eq_({k: opts[k] for k in opts if opts[k]}, {})
opts = insp.get_foreign_keys("user")[0]["options"]
eq_(opts, expected)
# eq_(dict((k, opts[k]) for k in opts if opts[k]), expected)
class NormalizedNameTest(fixtures.TablesTest):
__requires__ = ("denormalized_names",)
__backend__ = True
@classmethod
def define_tables(cls, metadata):
Table(
quoted_name("t1", quote=True),
metadata,
Column("id", Integer, primary_key=True),
)
Table(
quoted_name("t2", quote=True),
metadata,
Column("id", Integer, primary_key=True),
Column("t1id", ForeignKey("t1.id")),
)
def test_reflect_lowercase_forced_tables(self):
m2 = MetaData()
t2_ref = Table(
quoted_name("t2", quote=True), m2, autoload_with=config.db
)
t1_ref = m2.tables["t1"]
assert t2_ref.c.t1id.references(t1_ref.c.id)
m3 = MetaData()
m3.reflect(
config.db, only=lambda name, m: name.lower() in ("t1", "t2")
)
assert m3.tables["t2"].c.t1id.references(m3.tables["t1"].c.id)
def test_get_table_names(self):
tablenames = [
t
for t in inspect(config.db).get_table_names()
if t.lower() in ("t1", "t2")
]
eq_(tablenames[0].upper(), tablenames[0].lower())
eq_(tablenames[1].upper(), tablenames[1].lower())
class ComputedReflectionTest(fixtures.ComputedReflectionFixtureTest):
def test_computed_col_default_not_set(self):
insp = inspect(config.db)
cols = insp.get_columns("computed_default_table")
col_data = {c["name"]: c for c in cols}
is_true("42" in col_data["with_default"]["default"])
is_(col_data["normal"]["default"], None)
is_(col_data["computed_col"]["default"], None)
def test_get_column_returns_computed(self):
insp = inspect(config.db)
cols = insp.get_columns("computed_default_table")
data = {c["name"]: c for c in cols}
for key in ("id", "normal", "with_default"):
is_true("computed" not in data[key])
compData = data["computed_col"]
is_true("computed" in compData)
is_true("sqltext" in compData["computed"])
eq_(self.normalize(compData["computed"]["sqltext"]), "normal+42")
eq_(
"persisted" in compData["computed"],
testing.requires.computed_columns_reflect_persisted.enabled,
)
if testing.requires.computed_columns_reflect_persisted.enabled:
eq_(
compData["computed"]["persisted"],
testing.requires.computed_columns_default_persisted.enabled,
)
def check_column(self, data, column, sqltext, persisted):
is_true("computed" in data[column])
compData = data[column]["computed"]
eq_(self.normalize(compData["sqltext"]), sqltext)
if testing.requires.computed_columns_reflect_persisted.enabled:
is_true("persisted" in compData)
is_(compData["persisted"], persisted)
def test_get_column_returns_persisted(self):
insp = inspect(config.db)
cols = insp.get_columns("computed_column_table")
data = {c["name"]: c for c in cols}
self.check_column(
data,
"computed_no_flag",
"normal+42",
testing.requires.computed_columns_default_persisted.enabled,
)
if testing.requires.computed_columns_virtual.enabled:
self.check_column(
data,
"computed_virtual",
"normal+2",
False,
)
if testing.requires.computed_columns_stored.enabled:
self.check_column(
data,
"computed_stored",
"normal-42",
True,
)
@testing.requires.schemas
def test_get_column_returns_persisted_with_schema(self):
insp = inspect(config.db)
cols = insp.get_columns(
"computed_column_table", schema=config.test_schema
)
data = {c["name"]: c for c in cols}
self.check_column(
data,
"computed_no_flag",
"normal/42",
testing.requires.computed_columns_default_persisted.enabled,
)
if testing.requires.computed_columns_virtual.enabled:
self.check_column(
data,
"computed_virtual",
"normal/2",
False,
)
if testing.requires.computed_columns_stored.enabled:
self.check_column(
data,
"computed_stored",
"normal*42",
True,
)
class IdentityReflectionTest(fixtures.TablesTest):
run_inserts = run_deletes = None
__backend__ = True
__requires__ = ("identity_columns", "table_reflection")
@classmethod
def define_tables(cls, metadata):
Table(
"t1",
metadata,
Column("normal", Integer),
Column("id1", Integer, Identity()),
)
Table(
"t2",
metadata,
Column(
"id2",
Integer,
Identity(
always=True,
start=2,
increment=3,
minvalue=-2,
maxvalue=42,
cycle=True,
cache=4,
),
),
)
if testing.requires.schemas.enabled:
Table(
"t1",
metadata,
Column("normal", Integer),
Column("id1", Integer, Identity(always=True, start=20)),
schema=config.test_schema,
)
def check(self, value, exp, approx):
if testing.requires.identity_columns_standard.enabled:
common_keys = (
"always",
"start",
"increment",
"minvalue",
"maxvalue",
"cycle",
"cache",
)
for k in list(value):
if k not in common_keys:
value.pop(k)
if approx:
eq_(len(value), len(exp))
for k in value:
if k == "minvalue":
is_true(value[k] <= exp[k])
elif k in {"maxvalue", "cache"}:
is_true(value[k] >= exp[k])
else:
eq_(value[k], exp[k], k)
else:
eq_(value, exp)
else:
eq_(value["start"], exp["start"])
eq_(value["increment"], exp["increment"])
def test_reflect_identity(self):
insp = inspect(config.db)
cols = insp.get_columns("t1") + insp.get_columns("t2")
for col in cols:
if col["name"] == "normal":
is_false("identity" in col)
elif col["name"] == "id1":
if "autoincrement" in col:
is_true(col["autoincrement"])
eq_(col["default"], None)
is_true("identity" in col)
self.check(
col["identity"],
dict(
always=False,
start=1,
increment=1,
minvalue=1,
maxvalue=2147483647,
cycle=False,
cache=1,
),
approx=True,
)
elif col["name"] == "id2":
if "autoincrement" in col:
is_true(col["autoincrement"])
eq_(col["default"], None)
is_true("identity" in col)
self.check(
col["identity"],
dict(
always=True,
start=2,
increment=3,
minvalue=-2,
maxvalue=42,
cycle=True,
cache=4,
),
approx=False,
)
@testing.requires.schemas
def test_reflect_identity_schema(self):
insp = inspect(config.db)
cols = insp.get_columns("t1", schema=config.test_schema)
for col in cols:
if col["name"] == "normal":
is_false("identity" in col)
elif col["name"] == "id1":
if "autoincrement" in col:
is_true(col["autoincrement"])
eq_(col["default"], None)
is_true("identity" in col)
self.check(
col["identity"],
dict(
always=True,
start=20,
increment=1,
minvalue=1,
maxvalue=2147483647,
cycle=False,
cache=1,
),
approx=True,
)
class CompositeKeyReflectionTest(fixtures.TablesTest):
__backend__ = True
@classmethod
def define_tables(cls, metadata):
tb1 = Table(
"tb1",
metadata,
Column("id", Integer),
Column("attr", Integer),
Column("name", sql_types.VARCHAR(20)),
sa.PrimaryKeyConstraint("name", "id", "attr", name="pk_tb1"),
schema=None,
test_needs_fk=True,
)
Table(
"tb2",
metadata,
Column("id", Integer, primary_key=True),
Column("pid", Integer),
Column("pattr", Integer),
Column("pname", sql_types.VARCHAR(20)),
sa.ForeignKeyConstraint(
["pname", "pid", "pattr"],
[tb1.c.name, tb1.c.id, tb1.c.attr],
name="fk_tb1_name_id_attr",
),
schema=None,
test_needs_fk=True,
)
@testing.requires.primary_key_constraint_reflection
def test_pk_column_order(self, connection):
# test for issue #5661
insp = inspect(connection)
primary_key = insp.get_pk_constraint(self.tables.tb1.name)
eq_(primary_key.get("constrained_columns"), ["name", "id", "attr"])
@testing.requires.foreign_key_constraint_reflection
def test_fk_column_order(self, connection):
# test for issue #5661
insp = inspect(connection)
foreign_keys = insp.get_foreign_keys(self.tables.tb2.name)
eq_(len(foreign_keys), 1)
fkey1 = foreign_keys[0]
eq_(fkey1.get("referred_columns"), ["name", "id", "attr"])
eq_(fkey1.get("constrained_columns"), ["pname", "pid", "pattr"])
__all__ = (
"ComponentReflectionTest",
"ComponentReflectionTestExtra",
"TableNoColumnsTest",
"QuotedNameArgumentTest",
"BizarroCharacterFKResolutionTest",
"HasTableTest",
"HasIndexTest",
"NormalizedNameTest",
"ComputedReflectionTest",
"IdentityReflectionTest",
"CompositeKeyReflectionTest",
)