# sql/type_api.py
# Copyright (C) 2005-2024 the SQLAlchemy authors and contributors
# <see AUTHORS file>
#
# This module is part of SQLAlchemy and is released under
# the MIT License: https://www.opensource.org/licenses/mit-license.php
"""Base types API.
"""
from __future__ import annotations
from enum import Enum
from types import ModuleType
import typing
from typing import Any
from typing import Callable
from typing import cast
from typing import Dict
from typing import Generic
from typing import Mapping
from typing import NewType
from typing import Optional
from typing import overload
from typing import Sequence
from typing import Tuple
from typing import Type
from typing import TYPE_CHECKING
from typing import TypeVar
from typing import Union
from .base import SchemaEventTarget
from .cache_key import CacheConst
from .cache_key import NO_CACHE
from .operators import ColumnOperators
from .visitors import Visitable
from .. import exc
from .. import util
from ..util.typing import Protocol
from ..util.typing import Self
from ..util.typing import TypeAliasType
from ..util.typing import TypedDict
from ..util.typing import TypeGuard
# these are back-assigned by sqltypes.
if typing.TYPE_CHECKING:
from ._typing import _TypeEngineArgument
from .elements import BindParameter
from .elements import ColumnElement
from .operators import OperatorType
from .sqltypes import _resolve_value_to_type as _resolve_value_to_type
from .sqltypes import BOOLEANTYPE as BOOLEANTYPE # noqa: F401
from .sqltypes import INDEXABLE as INDEXABLE # noqa: F401
from .sqltypes import INTEGERTYPE as INTEGERTYPE # noqa: F401
from .sqltypes import MATCHTYPE as MATCHTYPE # noqa: F401
from .sqltypes import NULLTYPE as NULLTYPE
from .sqltypes import NUMERICTYPE as NUMERICTYPE # noqa: F401
from .sqltypes import STRINGTYPE as STRINGTYPE # noqa: F401
from .sqltypes import TABLEVALUE as TABLEVALUE # noqa: F401
from ..engine.interfaces import Dialect
from ..util.typing import GenericProtocol
_T = TypeVar("_T", bound=Any)
_T_co = TypeVar("_T_co", bound=Any, covariant=True)
_T_con = TypeVar("_T_con", bound=Any, contravariant=True)
_O = TypeVar("_O", bound=object)
_TE = TypeVar("_TE", bound="TypeEngine[Any]")
_CT = TypeVar("_CT", bound=Any)
_MatchedOnType = Union[
"GenericProtocol[Any]", TypeAliasType, NewType, Type[Any]
]
class _NoValueInList(Enum):
NO_VALUE_IN_LIST = 0
"""indicates we are trying to determine the type of an expression
against an empty list."""
_NO_VALUE_IN_LIST = _NoValueInList.NO_VALUE_IN_LIST
class _LiteralProcessorType(Protocol[_T_co]):
def __call__(self, value: Any) -> str: ...
class _BindProcessorType(Protocol[_T_con]):
def __call__(self, value: Optional[_T_con]) -> Any: ...
class _ResultProcessorType(Protocol[_T_co]):
def __call__(self, value: Any) -> Optional[_T_co]: ...
class _SentinelProcessorType(Protocol[_T_co]):
def __call__(self, value: Any) -> Optional[_T_co]: ...
class _BaseTypeMemoDict(TypedDict):
impl: TypeEngine[Any]
result: Dict[Any, Optional[_ResultProcessorType[Any]]]
class _TypeMemoDict(_BaseTypeMemoDict, total=False):
literal: Optional[_LiteralProcessorType[Any]]
bind: Optional[_BindProcessorType[Any]]
sentinel: Optional[_SentinelProcessorType[Any]]
custom: Dict[Any, object]
class _ComparatorFactory(Protocol[_T]):
def __call__(
self, expr: ColumnElement[_T]
) -> TypeEngine.Comparator[_T]: ...
class TypeEngine(Visitable, Generic[_T]):
"""The ultimate base class for all SQL datatypes.
Common subclasses of :class:`.TypeEngine` include
:class:`.String`, :class:`.Integer`, and :class:`.Boolean`.
For an overview of the SQLAlchemy typing system, see
:ref:`types_toplevel`.
.. seealso::
:ref:`types_toplevel`
"""
_sqla_type = True
_isnull = False
_is_tuple_type = False
_is_table_value = False
_is_array = False
_is_type_decorator = False
render_bind_cast = False
"""Render bind casts for :attr:`.BindTyping.RENDER_CASTS` mode.
If True, this type (usually a dialect level impl type) signals
to the compiler that a cast should be rendered around a bound parameter
for this type.
.. versionadded:: 2.0
.. seealso::
:class:`.BindTyping`
"""
render_literal_cast = False
"""render casts when rendering a value as an inline literal,
e.g. with :meth:`.TypeEngine.literal_processor`.
.. versionadded:: 2.0
"""
class Comparator(
ColumnOperators,
Generic[_CT],
):
"""Base class for custom comparison operations defined at the
type level. See :attr:`.TypeEngine.comparator_factory`.
"""
__slots__ = "expr", "type"
expr: ColumnElement[_CT]
type: TypeEngine[_CT]
def __clause_element__(self) -> ColumnElement[_CT]:
return self.expr
def __init__(self, expr: ColumnElement[_CT]):
self.expr = expr
self.type = expr.type
@util.preload_module("sqlalchemy.sql.default_comparator")
def operate(
self, op: OperatorType, *other: Any, **kwargs: Any
) -> ColumnElement[_CT]:
default_comparator = util.preloaded.sql_default_comparator
op_fn, addtl_kw = default_comparator.operator_lookup[op.__name__]
if kwargs:
addtl_kw = addtl_kw.union(kwargs)
return op_fn(self.expr, op, *other, **addtl_kw)
@util.preload_module("sqlalchemy.sql.default_comparator")
def reverse_operate(
self, op: OperatorType, other: Any, **kwargs: Any
) -> ColumnElement[_CT]:
default_comparator = util.preloaded.sql_default_comparator
op_fn, addtl_kw = default_comparator.operator_lookup[op.__name__]
if kwargs:
addtl_kw = addtl_kw.union(kwargs)
return op_fn(self.expr, op, other, reverse=True, **addtl_kw)
def _adapt_expression(
self,
op: OperatorType,
other_comparator: TypeEngine.Comparator[Any],
) -> Tuple[OperatorType, TypeEngine[Any]]:
"""evaluate the return type of <self> <op> <othertype>,
and apply any adaptations to the given operator.
This method determines the type of a resulting binary expression
given two source types and an operator. For example, two
:class:`_schema.Column` objects, both of the type
:class:`.Integer`, will
produce a :class:`.BinaryExpression` that also has the type
:class:`.Integer` when compared via the addition (``+``) operator.
However, using the addition operator with an :class:`.Integer`
and a :class:`.Date` object will produce a :class:`.Date`, assuming
"days delta" behavior by the database (in reality, most databases
other than PostgreSQL don't accept this particular operation).
The method returns a tuple of the form <operator>, <type>.
The resulting operator and type will be those applied to the
resulting :class:`.BinaryExpression` as the final operator and the
right-hand side of the expression.
Note that only a subset of operators make usage of
:meth:`._adapt_expression`,
including math operators and user-defined operators, but not
boolean comparison or special SQL keywords like MATCH or BETWEEN.
"""
return op, self.type
hashable = True
"""Flag, if False, means values from this type aren't hashable.
Used by the ORM when uniquing result lists.
"""
comparator_factory: _ComparatorFactory[Any] = Comparator
"""A :class:`.TypeEngine.Comparator` class which will apply
to operations performed by owning :class:`_expression.ColumnElement`
objects.
The :attr:`.comparator_factory` attribute is a hook consulted by
the core expression system when column and SQL expression operations
are performed. When a :class:`.TypeEngine.Comparator` class is
associated with this attribute, it allows custom re-definition of
all existing operators, as well as definition of new operators.
Existing operators include those provided by Python operator overloading
such as :meth:`.operators.ColumnOperators.__add__` and
:meth:`.operators.ColumnOperators.__eq__`,
those provided as standard
attributes of :class:`.operators.ColumnOperators` such as
:meth:`.operators.ColumnOperators.like`
and :meth:`.operators.ColumnOperators.in_`.
Rudimentary usage of this hook is allowed through simple subclassing
of existing types, or alternatively by using :class:`.TypeDecorator`.
See the documentation section :ref:`types_operators` for examples.
"""
sort_key_function: Optional[Callable[[Any], Any]] = None
"""A sorting function that can be passed as the key to sorted.
The default value of ``None`` indicates that the values stored by
this type are self-sorting.
.. versionadded:: 1.3.8
"""
should_evaluate_none: bool = False
"""If True, the Python constant ``None`` is considered to be handled
explicitly by this type.
The ORM uses this flag to indicate that a positive value of ``None``
is passed to the column in an INSERT statement, rather than omitting
the column from the INSERT statement which has the effect of firing
off column-level defaults. It also allows types which have special
behavior for Python None, such as a JSON type, to indicate that
they'd like to handle the None value explicitly.
To set this flag on an existing type, use the
:meth:`.TypeEngine.evaluates_none` method.
.. seealso::
:meth:`.TypeEngine.evaluates_none`
"""
_variant_mapping: util.immutabledict[str, TypeEngine[Any]] = (
util.EMPTY_DICT
)
def evaluates_none(self) -> Self:
"""Return a copy of this type which has the
:attr:`.should_evaluate_none` flag set to True.
E.g.::
Table(
'some_table', metadata,
Column(
String(50).evaluates_none(),
nullable=True,
server_default='no value')
)
The ORM uses this flag to indicate that a positive value of ``None``
is passed to the column in an INSERT statement, rather than omitting
the column from the INSERT statement which has the effect of firing
off column-level defaults. It also allows for types which have
special behavior associated with the Python None value to indicate
that the value doesn't necessarily translate into SQL NULL; a
prime example of this is a JSON type which may wish to persist the
JSON value ``'null'``.
In all cases, the actual NULL SQL value can be always be
persisted in any column by using
the :obj:`_expression.null` SQL construct in an INSERT statement
or associated with an ORM-mapped attribute.
.. note::
The "evaluates none" flag does **not** apply to a value
of ``None`` passed to :paramref:`_schema.Column.default` or
:paramref:`_schema.Column.server_default`; in these cases,
``None``
still means "no default".
.. seealso::
:ref:`session_forcing_null` - in the ORM documentation
:paramref:`.postgresql.JSON.none_as_null` - PostgreSQL JSON
interaction with this flag.
:attr:`.TypeEngine.should_evaluate_none` - class-level flag
"""
typ = self.copy()
typ.should_evaluate_none = True
return typ
def copy(self, **kw: Any) -> Self:
return self.adapt(self.__class__)
def copy_value(self, value: Any) -> Any:
return value
def literal_processor(
self, dialect: Dialect
) -> Optional[_LiteralProcessorType[_T]]:
"""Return a conversion function for processing literal values that are
to be rendered directly without using binds.
This function is used when the compiler makes use of the
"literal_binds" flag, typically used in DDL generation as well
as in certain scenarios where backends don't accept bound parameters.
Returns a callable which will receive a literal Python value
as the sole positional argument and will return a string representation
to be rendered in a SQL statement.
.. note::
This method is only called relative to a **dialect specific type
object**, which is often **private to a dialect in use** and is not
the same type object as the public facing one, which means it's not
feasible to subclass a :class:`.types.TypeEngine` class in order to
provide an alternate :meth:`_types.TypeEngine.literal_processor`
method, unless subclassing the :class:`_types.UserDefinedType`
class explicitly.
To provide alternate behavior for
:meth:`_types.TypeEngine.literal_processor`, implement a
:class:`_types.TypeDecorator` class and provide an implementation
of :meth:`_types.TypeDecorator.process_literal_param`.
.. seealso::
:ref:`types_typedecorator`
"""
return None
def bind_processor(
self, dialect: Dialect
) -> Optional[_BindProcessorType[_T]]:
"""Return a conversion function for processing bind values.
Returns a callable which will receive a bind parameter value
as the sole positional argument and will return a value to
send to the DB-API.
If processing is not necessary, the method should return ``None``.
.. note::
This method is only called relative to a **dialect specific type
object**, which is often **private to a dialect in use** and is not
the same type object as the public facing one, which means it's not
feasible to subclass a :class:`.types.TypeEngine` class in order to
provide an alternate :meth:`_types.TypeEngine.bind_processor`
method, unless subclassing the :class:`_types.UserDefinedType`
class explicitly.
To provide alternate behavior for
:meth:`_types.TypeEngine.bind_processor`, implement a
:class:`_types.TypeDecorator` class and provide an implementation
of :meth:`_types.TypeDecorator.process_bind_param`.
.. seealso::
:ref:`types_typedecorator`
:param dialect: Dialect instance in use.
"""
return None
def result_processor(
self, dialect: Dialect, coltype: object
) -> Optional[_ResultProcessorType[_T]]:
"""Return a conversion function for processing result row values.
Returns a callable which will receive a result row column
value as the sole positional argument and will return a value
to return to the user.
If processing is not necessary, the method should return ``None``.
.. note::
This method is only called relative to a **dialect specific type
object**, which is often **private to a dialect in use** and is not
the same type object as the public facing one, which means it's not
feasible to subclass a :class:`.types.TypeEngine` class in order to
provide an alternate :meth:`_types.TypeEngine.result_processor`
method, unless subclassing the :class:`_types.UserDefinedType`
class explicitly.
To provide alternate behavior for
:meth:`_types.TypeEngine.result_processor`, implement a
:class:`_types.TypeDecorator` class and provide an implementation
of :meth:`_types.TypeDecorator.process_result_value`.
.. seealso::
:ref:`types_typedecorator`
:param dialect: Dialect instance in use.
:param coltype: DBAPI coltype argument received in cursor.description.
"""
return None
def column_expression(
self, colexpr: ColumnElement[_T]
) -> Optional[ColumnElement[_T]]:
"""Given a SELECT column expression, return a wrapping SQL expression.
This is typically a SQL function that wraps a column expression
as rendered in the columns clause of a SELECT statement.
It is used for special data types that require
columns to be wrapped in some special database function in order
to coerce the value before being sent back to the application.
It is the SQL analogue of the :meth:`.TypeEngine.result_processor`
method.
This method is called during the **SQL compilation** phase of a
statement, when rendering a SQL string. It is **not** called
against specific values.
.. note::
This method is only called relative to a **dialect specific type
object**, which is often **private to a dialect in use** and is not
the same type object as the public facing one, which means it's not
feasible to subclass a :class:`.types.TypeEngine` class in order to
provide an alternate :meth:`_types.TypeEngine.column_expression`
method, unless subclassing the :class:`_types.UserDefinedType`
class explicitly.
To provide alternate behavior for
:meth:`_types.TypeEngine.column_expression`, implement a
:class:`_types.TypeDecorator` class and provide an implementation
of :meth:`_types.TypeDecorator.column_expression`.
.. seealso::
:ref:`types_typedecorator`
.. seealso::
:ref:`types_sql_value_processing`
"""
return None
@util.memoized_property
def _has_column_expression(self) -> bool:
"""memoized boolean, check if column_expression is implemented.
Allows the method to be skipped for the vast majority of expression
types that don't use this feature.
"""
return (
self.__class__.column_expression.__code__
is not TypeEngine.column_expression.__code__
)
def bind_expression(
self, bindvalue: BindParameter[_T]
) -> Optional[ColumnElement[_T]]:
"""Given a bind value (i.e. a :class:`.BindParameter` instance),
return a SQL expression in its place.
This is typically a SQL function that wraps the existing bound
parameter within the statement. It is used for special data types
that require literals being wrapped in some special database function
in order to coerce an application-level value into a database-specific
format. It is the SQL analogue of the
:meth:`.TypeEngine.bind_processor` method.
This method is called during the **SQL compilation** phase of a
statement, when rendering a SQL string. It is **not** called
against specific values.
Note that this method, when implemented, should always return
the exact same structure, without any conditional logic, as it
may be used in an executemany() call against an arbitrary number
of bound parameter sets.
.. note::
This method is only called relative to a **dialect specific type
object**, which is often **private to a dialect in use** and is not
the same type object as the public facing one, which means it's not
feasible to subclass a :class:`.types.TypeEngine` class in order to
provide an alternate :meth:`_types.TypeEngine.bind_expression`
method, unless subclassing the :class:`_types.UserDefinedType`
class explicitly.
To provide alternate behavior for
:meth:`_types.TypeEngine.bind_expression`, implement a
:class:`_types.TypeDecorator` class and provide an implementation
of :meth:`_types.TypeDecorator.bind_expression`.
.. seealso::
:ref:`types_typedecorator`
.. seealso::
:ref:`types_sql_value_processing`
"""
return None
@util.memoized_property
def _has_bind_expression(self) -> bool:
"""memoized boolean, check if bind_expression is implemented.
Allows the method to be skipped for the vast majority of expression
types that don't use this feature.
"""
return util.method_is_overridden(self, TypeEngine.bind_expression)
@staticmethod
def _to_instance(cls_or_self: Union[Type[_TE], _TE]) -> _TE:
return to_instance(cls_or_self)
def compare_values(self, x: Any, y: Any) -> bool:
"""Compare two values for equality."""
return x == y # type: ignore[no-any-return]
def get_dbapi_type(self, dbapi: ModuleType) -> Optional[Any]:
"""Return the corresponding type object from the underlying DB-API, if
any.
This can be useful for calling ``setinputsizes()``, for example.
"""
return None
@property
def python_type(self) -> Type[Any]:
"""Return the Python type object expected to be returned
by instances of this type, if known.
Basically, for those types which enforce a return type,
or are known across the board to do such for all common
DBAPIs (like ``int`` for example), will return that type.
If a return type is not defined, raises
``NotImplementedError``.
Note that any type also accommodates NULL in SQL which
means you can also get back ``None`` from any type
in practice.
"""
raise NotImplementedError()
def with_variant(
self,
type_: _TypeEngineArgument[Any],
*dialect_names: str,
) -> Self:
r"""Produce a copy of this type object that will utilize the given
type when applied to the dialect of the given name.
e.g.::
from sqlalchemy.types import String
from sqlalchemy.dialects import mysql
string_type = String()
string_type = string_type.with_variant(
mysql.VARCHAR(collation='foo'), 'mysql', 'mariadb'
)
The variant mapping indicates that when this type is
interpreted by a specific dialect, it will instead be
transmuted into the given type, rather than using the
primary type.
.. versionchanged:: 2.0 the :meth:`_types.TypeEngine.with_variant`
method now works with a :class:`_types.TypeEngine` object "in
place", returning a copy of the original type rather than returning
a wrapping object; the ``Variant`` class is no longer used.
:param type\_: a :class:`.TypeEngine` that will be selected
as a variant from the originating type, when a dialect
of the given name is in use.
:param \*dialect_names: one or more base names of the dialect which
uses this type. (i.e. ``'postgresql'``, ``'mysql'``, etc.)
.. versionchanged:: 2.0 multiple dialect names can be specified
for one variant.
.. seealso::
:ref:`types_with_variant` - illustrates the use of
:meth:`_types.TypeEngine.with_variant`.
"""
if not dialect_names:
raise exc.ArgumentError("At least one dialect name is required")
for dialect_name in dialect_names:
if dialect_name in self._variant_mapping:
raise exc.ArgumentError(
f"Dialect {dialect_name!r} is already present in "
f"the mapping for this {self!r}"
)
new_type = self.copy()
type_ = to_instance(type_)
if type_._variant_mapping:
raise exc.ArgumentError(
"can't pass a type that already has variants as a "
"dialect-level type to with_variant()"
)
new_type._variant_mapping = self._variant_mapping.union(
{dialect_name: type_ for dialect_name in dialect_names}
)
return new_type
def _resolve_for_literal(self, value: Any) -> Self:
"""adjust this type given a literal Python value that will be
stored in a bound parameter.
Used exclusively by _resolve_value_to_type().
.. versionadded:: 1.4.30 or 2.0
TODO: this should be part of public API
.. seealso::
:meth:`.TypeEngine._resolve_for_python_type`
"""
return self
def _resolve_for_python_type(
self,
python_type: Type[Any],
matched_on: _MatchedOnType,
matched_on_flattened: Type[Any],
) -> Optional[Self]:
"""given a Python type (e.g. ``int``, ``str``, etc. ) return an
instance of this :class:`.TypeEngine` that's appropriate for this type.
An additional argument ``matched_on`` is passed, which indicates an
entry from the ``__mro__`` of the given ``python_type`` that more
specifically matches how the caller located this :class:`.TypeEngine`
object. Such as, if a lookup of some kind links the ``int`` Python
type to the :class:`.Integer` SQL type, and the original object
was some custom subclass of ``int`` such as ``MyInt(int)``, the
arguments passed would be ``(MyInt, int)``.
If the given Python type does not correspond to this
:class:`.TypeEngine`, or the Python type is otherwise ambiguous, the
method should return None.
For simple cases, the method checks that the ``python_type``
and ``matched_on`` types are the same (i.e. not a subclass), and
returns self; for all other cases, it returns ``None``.
The initial use case here is for the ORM to link user-defined
Python standard library ``enum.Enum`` classes to the SQLAlchemy
:class:`.Enum` SQL type when constructing ORM Declarative mappings.
:param python_type: the Python type we want to use
:param matched_on: the Python type that led us to choose this
particular :class:`.TypeEngine` class, which would be a supertype
of ``python_type``. By default, the request is rejected if
``python_type`` doesn't match ``matched_on`` (None is returned).
.. versionadded:: 2.0.0b4
TODO: this should be part of public API
.. seealso::
:meth:`.TypeEngine._resolve_for_literal`
"""
if python_type is not matched_on_flattened:
return None
return self
def _with_collation(self, collation: str) -> Self:
"""set up error handling for the collate expression"""
raise NotImplementedError("this datatype does not support collation")
@util.ro_memoized_property
def _type_affinity(self) -> Optional[Type[TypeEngine[_T]]]:
"""Return a rudimental 'affinity' value expressing the general class
of type."""
typ = None
for t in self.__class__.__mro__:
if t is TypeEngine or TypeEngineMixin in t.__bases__:
return typ
elif issubclass(t, TypeEngine):
typ = t
else:
return self.__class__
@util.ro_memoized_property
def _generic_type_affinity(
self,
) -> Type[TypeEngine[_T]]:
best_camelcase = None
best_uppercase = None
if not isinstance(self, TypeEngine):
return self.__class__
for t in self.__class__.__mro__:
if (
t.__module__
in (
"sqlalchemy.sql.sqltypes",
"sqlalchemy.sql.type_api",
)
and issubclass(t, TypeEngine)
and TypeEngineMixin not in t.__bases__
and t not in (TypeEngine, TypeEngineMixin)
and t.__name__[0] != "_"
):
if t.__name__.isupper() and not best_uppercase:
best_uppercase = t
elif not t.__name__.isupper() and not best_camelcase:
best_camelcase = t
return (
best_camelcase
or best_uppercase
or cast("Type[TypeEngine[_T]]", NULLTYPE.__class__)
)
def as_generic(self, allow_nulltype: bool = False) -> TypeEngine[_T]:
"""
Return an instance of the generic type corresponding to this type
using heuristic rule. The method may be overridden if this
heuristic rule is not sufficient.
>>> from sqlalchemy.dialects.mysql import INTEGER
>>> INTEGER(display_width=4).as_generic()
Integer()
>>> from sqlalchemy.dialects.mysql import NVARCHAR
>>> NVARCHAR(length=100).as_generic()
Unicode(length=100)
.. versionadded:: 1.4.0b2
.. seealso::
:ref:`metadata_reflection_dbagnostic_types` - describes the
use of :meth:`_types.TypeEngine.as_generic` in conjunction with
the :meth:`_sql.DDLEvents.column_reflect` event, which is its
intended use.
"""
if (
not allow_nulltype
and self._generic_type_affinity == NULLTYPE.__class__
):
raise NotImplementedError(
"Default TypeEngine.as_generic() "
"heuristic method was unsuccessful for {}. A custom "
"as_generic() method must be implemented for this "
"type class.".format(
self.__class__.__module__ + "." + self.__class__.__name__
)
)
return util.constructor_copy(self, self._generic_type_affinity)
def dialect_impl(self, dialect: Dialect) -> TypeEngine[_T]:
"""Return a dialect-specific implementation for this
:class:`.TypeEngine`.
"""
try:
tm = dialect._type_memos[self]
except KeyError:
pass
else:
return tm["impl"]
return self._dialect_info(dialect)["impl"]
def _unwrapped_dialect_impl(self, dialect: Dialect) -> TypeEngine[_T]:
"""Return the 'unwrapped' dialect impl for this type.
For a type that applies wrapping logic (e.g. TypeDecorator), give
us the real, actual dialect-level type that is used.
This is used by TypeDecorator itself as well at least one case where
dialects need to check that a particular specific dialect-level
type is in use, within the :meth:`.DefaultDialect.set_input_sizes`
method.
"""
return self.dialect_impl(dialect)
def _cached_literal_processor(
self, dialect: Dialect
) -> Optional[_LiteralProcessorType[_T]]:
"""Return a dialect-specific literal processor for this type."""
try:
return dialect._type_memos[self]["literal"]
except KeyError:
pass
# avoid KeyError context coming into literal_processor() function
# raises
d = self._dialect_info(dialect)
d["literal"] = lp = d["impl"].literal_processor(dialect)
return lp
def _cached_bind_processor(
self, dialect: Dialect
) -> Optional[_BindProcessorType[_T]]:
"""Return a dialect-specific bind processor for this type."""
try:
return dialect._type_memos[self]["bind"]
except KeyError:
pass
# avoid KeyError context coming into bind_processor() function
# raises
d = self._dialect_info(dialect)
d["bind"] = bp = d["impl"].bind_processor(dialect)
return bp
def _cached_result_processor(
self, dialect: Dialect, coltype: Any
) -> Optional[_ResultProcessorType[_T]]:
"""Return a dialect-specific result processor for this type."""
try:
return dialect._type_memos[self]["result"][coltype]
except KeyError:
pass
# avoid KeyError context coming into result_processor() function
# raises
d = self._dialect_info(dialect)
# key assumption: DBAPI type codes are
# constants. Else this dictionary would
# grow unbounded.
rp = d["impl"].result_processor(dialect, coltype)
d["result"][coltype] = rp
return rp
def _cached_custom_processor(
self, dialect: Dialect, key: str, fn: Callable[[TypeEngine[_T]], _O]
) -> _O:
"""return a dialect-specific processing object for
custom purposes.
The cx_Oracle dialect uses this at the moment.
"""
try:
return cast(_O, dialect._type_memos[self]["custom"][key])
except KeyError:
pass
# avoid KeyError context coming into fn() function
# raises
d = self._dialect_info(dialect)
impl = d["impl"]
custom_dict = d.setdefault("custom", {})
custom_dict[key] = result = fn(impl)
return result
def _dialect_info(self, dialect: Dialect) -> _TypeMemoDict:
"""Return a dialect-specific registry which
caches a dialect-specific implementation, bind processing
function, and one or more result processing functions."""
if self in dialect._type_memos:
return dialect._type_memos[self]
else:
impl = self._gen_dialect_impl(dialect)
if impl is self:
impl = self.adapt(type(self))
# this can't be self, else we create a cycle
assert impl is not self
d: _TypeMemoDict = {"impl": impl, "result": {}}
dialect._type_memos[self] = d
return d
def _gen_dialect_impl(self, dialect: Dialect) -> TypeEngine[Any]:
if dialect.name in self._variant_mapping:
return self._variant_mapping[dialect.name]._gen_dialect_impl(
dialect
)
else:
return dialect.type_descriptor(self)
@util.memoized_property
def _static_cache_key(
self,
) -> Union[CacheConst, Tuple[Any, ...]]:
names = util.get_cls_kwargs(self.__class__)
return (self.__class__,) + tuple(
(
k,
(
self.__dict__[k]._static_cache_key
if isinstance(self.__dict__[k], TypeEngine)
else self.__dict__[k]
),
)
for k in names
if k in self.__dict__
and not k.startswith("_")
and self.__dict__[k] is not None
)
@overload
def adapt(self, cls: Type[_TE], **kw: Any) -> _TE: ...
@overload
def adapt(
self, cls: Type[TypeEngineMixin], **kw: Any
) -> TypeEngine[Any]: ...
def adapt(
self, cls: Type[Union[TypeEngine[Any], TypeEngineMixin]], **kw: Any
) -> TypeEngine[Any]:
"""Produce an "adapted" form of this type, given an "impl" class
to work with.
This method is used internally to associate generic
types with "implementation" types that are specific to a particular
dialect.
"""
typ = util.constructor_copy(
self, cast(Type[TypeEngine[Any]], cls), **kw
)
typ._variant_mapping = self._variant_mapping
return typ
def coerce_compared_value(
self, op: Optional[OperatorType], value: Any
) -> TypeEngine[Any]:
"""Suggest a type for a 'coerced' Python value in an expression.
Given an operator and value, gives the type a chance
to return a type which the value should be coerced into.
The default behavior here is conservative; if the right-hand
side is already coerced into a SQL type based on its
Python type, it is usually left alone.
End-user functionality extension here should generally be via
:class:`.TypeDecorator`, which provides more liberal behavior in that
it defaults to coercing the other side of the expression into this
type, thus applying special Python conversions above and beyond those
needed by the DBAPI to both ides. It also provides the public method
:meth:`.TypeDecorator.coerce_compared_value` which is intended for
end-user customization of this behavior.
"""
_coerced_type = _resolve_value_to_type(value)
if (
_coerced_type is NULLTYPE
or _coerced_type._type_affinity is self._type_affinity
):
return self
else:
return _coerced_type
def _compare_type_affinity(self, other: TypeEngine[Any]) -> bool:
return self._type_affinity is other._type_affinity
def compile(self, dialect: Optional[Dialect] = None) -> str:
"""Produce a string-compiled form of this :class:`.TypeEngine`.
When called with no arguments, uses a "default" dialect
to produce a string result.
:param dialect: a :class:`.Dialect` instance.
"""
# arg, return value is inconsistent with
# ClauseElement.compile()....this is a mistake.
if dialect is None:
dialect = self._default_dialect()
return dialect.type_compiler_instance.process(self)
@util.preload_module("sqlalchemy.engine.default")
def _default_dialect(self) -> Dialect:
default = util.preloaded.engine_default
# dmypy / mypy seems to sporadically keep thinking this line is
# returning Any, which seems to be caused by the @deprecated_params
# decorator on the DefaultDialect constructor
return default.StrCompileDialect() # type: ignore
def __str__(self) -> str:
return str(self.compile())
def __repr__(self) -> str:
return util.generic_repr(self)
class TypeEngineMixin:
"""classes which subclass this can act as "mixin" classes for
TypeEngine."""
__slots__ = ()
if TYPE_CHECKING:
@util.memoized_property
def _static_cache_key(
self,
) -> Union[CacheConst, Tuple[Any, ...]]: ...
@overload
def adapt(self, cls: Type[_TE], **kw: Any) -> _TE: ...
@overload
def adapt(
self, cls: Type[TypeEngineMixin], **kw: Any
) -> TypeEngine[Any]: ...
def adapt(
self, cls: Type[Union[TypeEngine[Any], TypeEngineMixin]], **kw: Any
) -> TypeEngine[Any]: ...
def dialect_impl(self, dialect: Dialect) -> TypeEngine[Any]: ...
class ExternalType(TypeEngineMixin):
"""mixin that defines attributes and behaviors specific to third-party
datatypes.
"Third party" refers to datatypes that are defined outside the scope
of SQLAlchemy within either end-user application code or within
external extensions to SQLAlchemy.
Subclasses currently include :class:`.TypeDecorator` and
:class:`.UserDefinedType`.
.. versionadded:: 1.4.28
"""
cache_ok: Optional[bool] = None
"""Indicate if statements using this :class:`.ExternalType` are "safe to
cache".
The default value ``None`` will emit a warning and then not allow caching
of a statement which includes this type. Set to ``False`` to disable
statements using this type from being cached at all without a warning.
When set to ``True``, the object's class and selected elements from its
state will be used as part of the cache key. For example, using a
:class:`.TypeDecorator`::
class MyType(TypeDecorator):
impl = String
cache_ok = True
def __init__(self, choices):
self.choices = tuple(choices)
self.internal_only = True
The cache key for the above type would be equivalent to::
>>> MyType(["a", "b", "c"])._static_cache_key
(<class '__main__.MyType'>, ('choices', ('a', 'b', 'c')))
The caching scheme will extract attributes from the type that correspond
to the names of parameters in the ``__init__()`` method. Above, the
"choices" attribute becomes part of the cache key but "internal_only"
does not, because there is no parameter named "internal_only".
The requirements for cacheable elements is that they are hashable
and also that they indicate the same SQL rendered for expressions using
this type every time for a given cache value.
To accommodate for datatypes that refer to unhashable structures such
as dictionaries, sets and lists, these objects can be made "cacheable"
by assigning hashable structures to the attributes whose names
correspond with the names of the arguments. For example, a datatype
which accepts a dictionary of lookup values may publish this as a sorted
series of tuples. Given a previously un-cacheable type as::
class LookupType(UserDefinedType):
'''a custom type that accepts a dictionary as a parameter.
this is the non-cacheable version, as "self.lookup" is not
hashable.
'''
def __init__(self, lookup):
self.lookup = lookup
def get_col_spec(self, **kw):
return "VARCHAR(255)"
def bind_processor(self, dialect):
# ... works with "self.lookup" ...
Where "lookup" is a dictionary. The type will not be able to generate
a cache key::
>>> type_ = LookupType({"a": 10, "b": 20})
>>> type_._static_cache_key
<stdin>:1: SAWarning: UserDefinedType LookupType({'a': 10, 'b': 20}) will not
produce a cache key because the ``cache_ok`` flag is not set to True.
Set this flag to True if this type object's state is safe to use
in a cache key, or False to disable this warning.
symbol('no_cache')
If we **did** set up such a cache key, it wouldn't be usable. We would
get a tuple structure that contains a dictionary inside of it, which
cannot itself be used as a key in a "cache dictionary" such as SQLAlchemy's
statement cache, since Python dictionaries aren't hashable::
>>> # set cache_ok = True
>>> type_.cache_ok = True
>>> # this is the cache key it would generate
>>> key = type_._static_cache_key
>>> key
(<class '__main__.LookupType'>, ('lookup', {'a': 10, 'b': 20}))
>>> # however this key is not hashable, will fail when used with
>>> # SQLAlchemy statement cache
>>> some_cache = {key: "some sql value"}
Traceback (most recent call last): File "<stdin>", line 1,
in <module> TypeError: unhashable type: 'dict'
The type may be made cacheable by assigning a sorted tuple of tuples
to the ".lookup" attribute::
class LookupType(UserDefinedType):
'''a custom type that accepts a dictionary as a parameter.
The dictionary is stored both as itself in a private variable,
and published in a public variable as a sorted tuple of tuples,
which is hashable and will also return the same value for any
two equivalent dictionaries. Note it assumes the keys and
values of the dictionary are themselves hashable.
'''
cache_ok = True
def __init__(self, lookup):
self._lookup = lookup
# assume keys/values of "lookup" are hashable; otherwise
# they would also need to be converted in some way here
self.lookup = tuple(
(key, lookup[key]) for key in sorted(lookup)
)
def get_col_spec(self, **kw):
return "VARCHAR(255)"
def bind_processor(self, dialect):
# ... works with "self._lookup" ...
Where above, the cache key for ``LookupType({"a": 10, "b": 20})`` will be::
>>> LookupType({"a": 10, "b": 20})._static_cache_key
(<class '__main__.LookupType'>, ('lookup', (('a', 10), ('b', 20))))
.. versionadded:: 1.4.14 - added the ``cache_ok`` flag to allow
some configurability of caching for :class:`.TypeDecorator` classes.
.. versionadded:: 1.4.28 - added the :class:`.ExternalType` mixin which
generalizes the ``cache_ok`` flag to both the :class:`.TypeDecorator`
and :class:`.UserDefinedType` classes.
.. seealso::
:ref:`sql_caching`
""" # noqa: E501
@util.non_memoized_property
def _static_cache_key(
self,
) -> Union[CacheConst, Tuple[Any, ...]]:
cache_ok = self.__class__.__dict__.get("cache_ok", None)
if cache_ok is None:
for subtype in self.__class__.__mro__:
if ExternalType in subtype.__bases__:
break
else:
subtype = self.__class__.__mro__[1]
util.warn(
"%s %r will not produce a cache key because "
"the ``cache_ok`` attribute is not set to True. This can "
"have significant performance implications including some "
"performance degradations in comparison to prior SQLAlchemy "
"versions. Set this attribute to True if this type object's "
"state is safe to use in a cache key, or False to "
"disable this warning." % (subtype.__name__, self),
code="cprf",
)
elif cache_ok is True:
return super()._static_cache_key
return NO_CACHE
class UserDefinedType(
ExternalType, TypeEngineMixin, TypeEngine[_T], util.EnsureKWArg
):
"""Base for user defined types.
This should be the base of new types. Note that
for most cases, :class:`.TypeDecorator` is probably
more appropriate::
import sqlalchemy.types as types
class MyType(types.UserDefinedType):
cache_ok = True
def __init__(self, precision = 8):
self.precision = precision
def get_col_spec(self, **kw):
return "MYTYPE(%s)" % self.precision
def bind_processor(self, dialect):
def process(value):
return value
return process
def result_processor(self, dialect, coltype):
def process(value):
return value
return process
Once the type is made, it's immediately usable::
table = Table('foo', metadata_obj,
Column('id', Integer, primary_key=True),
Column('data', MyType(16))
)
The ``get_col_spec()`` method will in most cases receive a keyword
argument ``type_expression`` which refers to the owning expression
of the type as being compiled, such as a :class:`_schema.Column` or
:func:`.cast` construct. This keyword is only sent if the method
accepts keyword arguments (e.g. ``**kw``) in its argument signature;
introspection is used to check for this in order to support legacy
forms of this function.
The :attr:`.UserDefinedType.cache_ok` class-level flag indicates if this
custom :class:`.UserDefinedType` is safe to be used as part of a cache key.
This flag defaults to ``None`` which will initially generate a warning
when the SQL compiler attempts to generate a cache key for a statement
that uses this type. If the :class:`.UserDefinedType` is not guaranteed
to produce the same bind/result behavior and SQL generation
every time, this flag should be set to ``False``; otherwise if the
class produces the same behavior each time, it may be set to ``True``.
See :attr:`.UserDefinedType.cache_ok` for further notes on how this works.
.. versionadded:: 1.4.28 Generalized the :attr:`.ExternalType.cache_ok`
flag so that it is available for both :class:`.TypeDecorator` as well
as :class:`.UserDefinedType`.
"""
__visit_name__ = "user_defined"
ensure_kwarg = "get_col_spec"
def coerce_compared_value(
self, op: Optional[OperatorType], value: Any
) -> TypeEngine[Any]:
"""Suggest a type for a 'coerced' Python value in an expression.
Default behavior for :class:`.UserDefinedType` is the
same as that of :class:`.TypeDecorator`; by default it returns
``self``, assuming the compared value should be coerced into
the same type as this one. See
:meth:`.TypeDecorator.coerce_compared_value` for more detail.
"""
return self
class Emulated(TypeEngineMixin):
"""Mixin for base types that emulate the behavior of a DB-native type.
An :class:`.Emulated` type will use an available database type
in conjunction with Python-side routines and/or database constraints
in order to approximate the behavior of a database type that is provided
natively by some backends. When a native-providing backend is in
use, the native version of the type is used. This native version
should include the :class:`.NativeForEmulated` mixin to allow it to be
distinguished from :class:`.Emulated`.
Current examples of :class:`.Emulated` are: :class:`.Interval`,
:class:`.Enum`, :class:`.Boolean`.
.. versionadded:: 1.2.0b3
"""
native: bool
def adapt_to_emulated(
self,
impltype: Type[Union[TypeEngine[Any], TypeEngineMixin]],
**kw: Any,
) -> TypeEngine[Any]:
"""Given an impl class, adapt this type to the impl assuming
"emulated".
The impl should also be an "emulated" version of this type,
most likely the same class as this type itself.
e.g.: sqltypes.Enum adapts to the Enum class.
"""
return super().adapt(impltype, **kw)
@overload
def adapt(self, cls: Type[_TE], **kw: Any) -> _TE: ...
@overload
def adapt(
self, cls: Type[TypeEngineMixin], **kw: Any
) -> TypeEngine[Any]: ...
def adapt(
self, cls: Type[Union[TypeEngine[Any], TypeEngineMixin]], **kw: Any
) -> TypeEngine[Any]:
if _is_native_for_emulated(cls):
if self.native:
# native support requested, dialect gave us a native
# implementor, pass control over to it
return cls.adapt_emulated_to_native(self, **kw)
else:
# non-native support, let the native implementor
# decide also, at the moment this is just to help debugging
# as only the default logic is implemented.
return cls.adapt_native_to_emulated(self, **kw)
else:
# this would be, both classes are Enum, or both classes
# are postgresql.ENUM
if issubclass(cls, self.__class__):
return self.adapt_to_emulated(cls, **kw)
else:
return super().adapt(cls, **kw)
def _is_native_for_emulated(
typ: Type[Union[TypeEngine[Any], TypeEngineMixin]],
) -> TypeGuard[Type[NativeForEmulated]]:
return hasattr(typ, "adapt_emulated_to_native")
class NativeForEmulated(TypeEngineMixin):
"""Indicates DB-native types supported by an :class:`.Emulated` type.
.. versionadded:: 1.2.0b3
"""
@classmethod
def adapt_native_to_emulated(
cls,
impl: Union[TypeEngine[Any], TypeEngineMixin],
**kw: Any,
) -> TypeEngine[Any]:
"""Given an impl, adapt this type's class to the impl assuming
"emulated".
"""
impltype = impl.__class__
return impl.adapt(impltype, **kw)
@classmethod
def adapt_emulated_to_native(
cls,
impl: Union[TypeEngine[Any], TypeEngineMixin],
**kw: Any,
) -> TypeEngine[Any]:
"""Given an impl, adapt this type's class to the impl assuming
"native".
The impl will be an :class:`.Emulated` class but not a
:class:`.NativeForEmulated`.
e.g.: postgresql.ENUM produces a type given an Enum instance.
"""
# dmypy seems to crash on this
return cls(**kw) # type: ignore
# dmypy seems to crash with this, on repeated runs with changes
# if TYPE_CHECKING:
# def __init__(self, **kw: Any):
# ...
class TypeDecorator(SchemaEventTarget, ExternalType, TypeEngine[_T]):
"""Allows the creation of types which add additional functionality
to an existing type.
This method is preferred to direct subclassing of SQLAlchemy's
built-in types as it ensures that all required functionality of
the underlying type is kept in place.
Typical usage::
import sqlalchemy.types as types
class MyType(types.TypeDecorator):
'''Prefixes Unicode values with "PREFIX:" on the way in and
strips it off on the way out.
'''
impl = types.Unicode
cache_ok = True
def process_bind_param(self, value, dialect):
return "PREFIX:" + value
def process_result_value(self, value, dialect):
return value[7:]
def copy(self, **kw):
return MyType(self.impl.length)
The class-level ``impl`` attribute is required, and can reference any
:class:`.TypeEngine` class. Alternatively, the :meth:`load_dialect_impl`
method can be used to provide different type classes based on the dialect
given; in this case, the ``impl`` variable can reference
``TypeEngine`` as a placeholder.
The :attr:`.TypeDecorator.cache_ok` class-level flag indicates if this
custom :class:`.TypeDecorator` is safe to be used as part of a cache key.
This flag defaults to ``None`` which will initially generate a warning
when the SQL compiler attempts to generate a cache key for a statement
that uses this type. If the :class:`.TypeDecorator` is not guaranteed
to produce the same bind/result behavior and SQL generation
every time, this flag should be set to ``False``; otherwise if the
class produces the same behavior each time, it may be set to ``True``.
See :attr:`.TypeDecorator.cache_ok` for further notes on how this works.
Types that receive a Python type that isn't similar to the ultimate type
used may want to define the :meth:`TypeDecorator.coerce_compared_value`
method. This is used to give the expression system a hint when coercing
Python objects into bind parameters within expressions. Consider this
expression::
mytable.c.somecol + datetime.date(2009, 5, 15)
Above, if "somecol" is an ``Integer`` variant, it makes sense that
we're doing date arithmetic, where above is usually interpreted
by databases as adding a number of days to the given date.
The expression system does the right thing by not attempting to
coerce the "date()" value into an integer-oriented bind parameter.
However, in the case of ``TypeDecorator``, we are usually changing an
incoming Python type to something new - ``TypeDecorator`` by default will
"coerce" the non-typed side to be the same type as itself. Such as below,
we define an "epoch" type that stores a date value as an integer::
class MyEpochType(types.TypeDecorator):
impl = types.Integer
cache_ok = True
epoch = datetime.date(1970, 1, 1)
def process_bind_param(self, value, dialect):
return (value - self.epoch).days
def process_result_value(self, value, dialect):
return self.epoch + timedelta(days=value)
Our expression of ``somecol + date`` with the above type will coerce the
"date" on the right side to also be treated as ``MyEpochType``.
This behavior can be overridden via the
:meth:`~TypeDecorator.coerce_compared_value` method, which returns a type
that should be used for the value of the expression. Below we set it such
that an integer value will be treated as an ``Integer``, and any other
value is assumed to be a date and will be treated as a ``MyEpochType``::
def coerce_compared_value(self, op, value):
if isinstance(value, int):
return Integer()
else:
return self
.. warning::
Note that the **behavior of coerce_compared_value is not inherited
by default from that of the base type**.
If the :class:`.TypeDecorator` is augmenting a
type that requires special logic for certain types of operators,
this method **must** be overridden. A key example is when decorating
the :class:`_postgresql.JSON` and :class:`_postgresql.JSONB` types;
the default rules of :meth:`.TypeEngine.coerce_compared_value` should
be used in order to deal with operators like index operations::
from sqlalchemy import JSON
from sqlalchemy import TypeDecorator
class MyJsonType(TypeDecorator):
impl = JSON
cache_ok = True
def coerce_compared_value(self, op, value):
return self.impl.coerce_compared_value(op, value)
Without the above step, index operations such as ``mycol['foo']``
will cause the index value ``'foo'`` to be JSON encoded.
Similarly, when working with the :class:`.ARRAY` datatype, the
type coercion for index operations (e.g. ``mycol[5]``) is also
handled by :meth:`.TypeDecorator.coerce_compared_value`, where
again a simple override is sufficient unless special rules are needed
for particular operators::
from sqlalchemy import ARRAY
from sqlalchemy import TypeDecorator
class MyArrayType(TypeDecorator):
impl = ARRAY
cache_ok = True
def coerce_compared_value(self, op, value):
return self.impl.coerce_compared_value(op, value)
"""
__visit_name__ = "type_decorator"
_is_type_decorator = True
# this is that pattern I've used in a few places (Dialect.dbapi,
# Dialect.type_compiler) where the "cls.attr" is a class to make something,
# and "instance.attr" is an instance of that thing. It's such a nifty,
# great pattern, and there is zero chance Python typing tools will ever be
# OK with it. For TypeDecorator.impl, this is a highly public attribute so
# we really can't change its behavior without a major deprecation routine.
impl: Union[TypeEngine[Any], Type[TypeEngine[Any]]]
# we are changing its behavior *slightly*, which is that we now consume
# the instance level version from this memoized property instead, so you
# can't reassign "impl" on an existing TypeDecorator that's already been
# used (something one shouldn't do anyway) without also updating
# impl_instance.
@util.memoized_property
def impl_instance(self) -> TypeEngine[Any]:
return self.impl # type: ignore
def __init__(self, *args: Any, **kwargs: Any):
"""Construct a :class:`.TypeDecorator`.
Arguments sent here are passed to the constructor
of the class assigned to the ``impl`` class level attribute,
assuming the ``impl`` is a callable, and the resulting
object is assigned to the ``self.impl`` instance attribute
(thus overriding the class attribute of the same name).
If the class level ``impl`` is not a callable (the unusual case),
it will be assigned to the same instance attribute 'as-is',
ignoring those arguments passed to the constructor.
Subclasses can override this to customize the generation
of ``self.impl`` entirely.
"""
if not hasattr(self.__class__, "impl"):
raise AssertionError(
"TypeDecorator implementations "
"require a class-level variable "
"'impl' which refers to the class of "
"type being decorated"
)
self.impl = to_instance(self.__class__.impl, *args, **kwargs)
coerce_to_is_types: Sequence[Type[Any]] = (type(None),)
"""Specify those Python types which should be coerced at the expression
level to "IS <constant>" when compared using ``==`` (and same for
``IS NOT`` in conjunction with ``!=``).
For most SQLAlchemy types, this includes ``NoneType``, as well as
``bool``.
:class:`.TypeDecorator` modifies this list to only include ``NoneType``,
as typedecorator implementations that deal with boolean types are common.
Custom :class:`.TypeDecorator` classes can override this attribute to
return an empty tuple, in which case no values will be coerced to
constants.
"""
class Comparator(TypeEngine.Comparator[_CT]):
"""A :class:`.TypeEngine.Comparator` that is specific to
:class:`.TypeDecorator`.
User-defined :class:`.TypeDecorator` classes should not typically
need to modify this.
"""
__slots__ = ()
def operate(
self, op: OperatorType, *other: Any, **kwargs: Any
) -> ColumnElement[_CT]:
if TYPE_CHECKING:
assert isinstance(self.expr.type, TypeDecorator)
kwargs["_python_is_types"] = self.expr.type.coerce_to_is_types
return super().operate(op, *other, **kwargs)
def reverse_operate(
self, op: OperatorType, other: Any, **kwargs: Any
) -> ColumnElement[_CT]:
if TYPE_CHECKING:
assert isinstance(self.expr.type, TypeDecorator)
kwargs["_python_is_types"] = self.expr.type.coerce_to_is_types
return super().reverse_operate(op, other, **kwargs)
@property
def comparator_factory( # type: ignore # mypy properties bug
self,
) -> _ComparatorFactory[Any]:
if TypeDecorator.Comparator in self.impl.comparator_factory.__mro__: # type: ignore # noqa: E501
return self.impl.comparator_factory
else:
# reconcile the Comparator class on the impl with that
# of TypeDecorator
return type(
"TDComparator",
(TypeDecorator.Comparator, self.impl.comparator_factory), # type: ignore # noqa: E501
{},
)
def _copy_with_check(self) -> Self:
tt = self.copy()
if not isinstance(tt, self.__class__):
raise AssertionError(
"Type object %s does not properly "
"implement the copy() method, it must "
"return an object of type %s" % (self, self.__class__)
)
return tt
def _gen_dialect_impl(self, dialect: Dialect) -> TypeEngine[_T]:
if dialect.name in self._variant_mapping:
adapted = dialect.type_descriptor(
self._variant_mapping[dialect.name]
)
else:
adapted = dialect.type_descriptor(self)
if adapted is not self:
return adapted
# otherwise adapt the impl type, link
# to a copy of this TypeDecorator and return
# that.
typedesc = self.load_dialect_impl(dialect).dialect_impl(dialect)
tt = self._copy_with_check()
tt.impl = tt.impl_instance = typedesc
return tt
def _with_collation(self, collation: str) -> Self:
tt = self._copy_with_check()
tt.impl = tt.impl_instance = self.impl_instance._with_collation(
collation
)
return tt
@util.ro_non_memoized_property
def _type_affinity(self) -> Optional[Type[TypeEngine[Any]]]:
return self.impl_instance._type_affinity
def _set_parent(
self, parent: SchemaEventTarget, outer: bool = False, **kw: Any
) -> None:
"""Support SchemaEventTarget"""
super()._set_parent(parent)
if not outer and isinstance(self.impl_instance, SchemaEventTarget):
self.impl_instance._set_parent(parent, outer=False, **kw)
def _set_parent_with_dispatch(
self, parent: SchemaEventTarget, **kw: Any
) -> None:
"""Support SchemaEventTarget"""
super()._set_parent_with_dispatch(parent, outer=True, **kw)
if isinstance(self.impl_instance, SchemaEventTarget):
self.impl_instance._set_parent_with_dispatch(parent)
def type_engine(self, dialect: Dialect) -> TypeEngine[Any]:
"""Return a dialect-specific :class:`.TypeEngine` instance
for this :class:`.TypeDecorator`.
In most cases this returns a dialect-adapted form of
the :class:`.TypeEngine` type represented by ``self.impl``.
Makes usage of :meth:`dialect_impl`.
Behavior can be customized here by overriding
:meth:`load_dialect_impl`.
"""
adapted = dialect.type_descriptor(self)
if not isinstance(adapted, type(self)):
return adapted
else:
return self.load_dialect_impl(dialect)
def load_dialect_impl(self, dialect: Dialect) -> TypeEngine[Any]:
"""Return a :class:`.TypeEngine` object corresponding to a dialect.
This is an end-user override hook that can be used to provide
differing types depending on the given dialect. It is used
by the :class:`.TypeDecorator` implementation of :meth:`type_engine`
to help determine what type should ultimately be returned
for a given :class:`.TypeDecorator`.
By default returns ``self.impl``.
"""
return self.impl_instance
def _unwrapped_dialect_impl(self, dialect: Dialect) -> TypeEngine[Any]:
"""Return the 'unwrapped' dialect impl for this type.
This is used by the :meth:`.DefaultDialect.set_input_sizes`
method.
"""
# some dialects have a lookup for a TypeDecorator subclass directly.
# postgresql.INTERVAL being the main example
typ = self.dialect_impl(dialect)
# if we are still a type decorator, load the per-dialect switch
# (such as what Variant uses), then get the dialect impl for that.
if isinstance(typ, self.__class__):
return typ.load_dialect_impl(dialect).dialect_impl(dialect)
else:
return typ
def __getattr__(self, key: str) -> Any:
"""Proxy all other undefined accessors to the underlying
implementation."""
return getattr(self.impl_instance, key)
def process_literal_param(
self, value: Optional[_T], dialect: Dialect
) -> str:
"""Receive a literal parameter value to be rendered inline within
a statement.
.. note::
This method is called during the **SQL compilation** phase of a
statement, when rendering a SQL string. Unlike other SQL
compilation methods, it is passed a specific Python value to be
rendered as a string. However it should not be confused with the
:meth:`_types.TypeDecorator.process_bind_param` method, which is
the more typical method that processes the actual value passed to a
particular parameter at statement execution time.
Custom subclasses of :class:`_types.TypeDecorator` should override
this method to provide custom behaviors for incoming data values
that are in the special case of being rendered as literals.
The returned string will be rendered into the output string.
"""
raise NotImplementedError()
def process_bind_param(self, value: Optional[_T], dialect: Dialect) -> Any:
"""Receive a bound parameter value to be converted.
Custom subclasses of :class:`_types.TypeDecorator` should override
this method to provide custom behaviors for incoming data values.
This method is called at **statement execution time** and is passed
the literal Python data value which is to be associated with a bound
parameter in the statement.
The operation could be anything desired to perform custom
behavior, such as transforming or serializing data.
This could also be used as a hook for validating logic.
:param value: Data to operate upon, of any type expected by
this method in the subclass. Can be ``None``.
:param dialect: the :class:`.Dialect` in use.
.. seealso::
:ref:`types_typedecorator`
:meth:`_types.TypeDecorator.process_result_value`
"""
raise NotImplementedError()
def process_result_value(
self, value: Optional[Any], dialect: Dialect
) -> Optional[_T]:
"""Receive a result-row column value to be converted.
Custom subclasses of :class:`_types.TypeDecorator` should override
this method to provide custom behaviors for data values
being received in result rows coming from the database.
This method is called at **result fetching time** and is passed
the literal Python data value that's extracted from a database result
row.
The operation could be anything desired to perform custom
behavior, such as transforming or deserializing data.
:param value: Data to operate upon, of any type expected by
this method in the subclass. Can be ``None``.
:param dialect: the :class:`.Dialect` in use.
.. seealso::
:ref:`types_typedecorator`
:meth:`_types.TypeDecorator.process_bind_param`
"""
raise NotImplementedError()
@util.memoized_property
def _has_bind_processor(self) -> bool:
"""memoized boolean, check if process_bind_param is implemented.
Allows the base process_bind_param to raise
NotImplementedError without needing to test an expensive
exception throw.
"""
return util.method_is_overridden(
self, TypeDecorator.process_bind_param
)
@util.memoized_property
def _has_literal_processor(self) -> bool:
"""memoized boolean, check if process_literal_param is implemented."""
return util.method_is_overridden(
self, TypeDecorator.process_literal_param
)
def literal_processor(
self, dialect: Dialect
) -> Optional[_LiteralProcessorType[_T]]:
"""Provide a literal processing function for the given
:class:`.Dialect`.
This is the method that fulfills the :class:`.TypeEngine`
contract for literal value conversion which normally occurs via
the :meth:`_types.TypeEngine.literal_processor` method.
.. note::
User-defined subclasses of :class:`_types.TypeDecorator` should
**not** implement this method, and should instead implement
:meth:`_types.TypeDecorator.process_literal_param` so that the
"inner" processing provided by the implementing type is maintained.
"""
if self._has_literal_processor:
process_literal_param = self.process_literal_param
process_bind_param = None
elif self._has_bind_processor:
# use the bind processor if dont have a literal processor,
# but we have an impl literal processor
process_literal_param = None
process_bind_param = self.process_bind_param
else:
process_literal_param = None
process_bind_param = None
if process_literal_param is not None:
impl_processor = self.impl_instance.literal_processor(dialect)
if impl_processor:
fixed_impl_processor = impl_processor
fixed_process_literal_param = process_literal_param
def process(value: Any) -> str:
return fixed_impl_processor(
fixed_process_literal_param(value, dialect)
)
else:
fixed_process_literal_param = process_literal_param
def process(value: Any) -> str:
return fixed_process_literal_param(value, dialect)
return process
elif process_bind_param is not None:
impl_processor = self.impl_instance.literal_processor(dialect)
if not impl_processor:
return None
else:
fixed_impl_processor = impl_processor
fixed_process_bind_param = process_bind_param
def process(value: Any) -> str:
return fixed_impl_processor(
fixed_process_bind_param(value, dialect)
)
return process
else:
return self.impl_instance.literal_processor(dialect)
def bind_processor(
self, dialect: Dialect
) -> Optional[_BindProcessorType[_T]]:
"""Provide a bound value processing function for the
given :class:`.Dialect`.
This is the method that fulfills the :class:`.TypeEngine`
contract for bound value conversion which normally occurs via
the :meth:`_types.TypeEngine.bind_processor` method.
.. note::
User-defined subclasses of :class:`_types.TypeDecorator` should
**not** implement this method, and should instead implement
:meth:`_types.TypeDecorator.process_bind_param` so that the "inner"
processing provided by the implementing type is maintained.
:param dialect: Dialect instance in use.
"""
if self._has_bind_processor:
process_param = self.process_bind_param
impl_processor = self.impl_instance.bind_processor(dialect)
if impl_processor:
fixed_impl_processor = impl_processor
fixed_process_param = process_param
def process(value: Optional[_T]) -> Any:
return fixed_impl_processor(
fixed_process_param(value, dialect)
)
else:
fixed_process_param = process_param
def process(value: Optional[_T]) -> Any:
return fixed_process_param(value, dialect)
return process
else:
return self.impl_instance.bind_processor(dialect)
@util.memoized_property
def _has_result_processor(self) -> bool:
"""memoized boolean, check if process_result_value is implemented.
Allows the base process_result_value to raise
NotImplementedError without needing to test an expensive
exception throw.
"""
return util.method_is_overridden(
self, TypeDecorator.process_result_value
)
def result_processor(
self, dialect: Dialect, coltype: Any
) -> Optional[_ResultProcessorType[_T]]:
"""Provide a result value processing function for the given
:class:`.Dialect`.
This is the method that fulfills the :class:`.TypeEngine`
contract for bound value conversion which normally occurs via
the :meth:`_types.TypeEngine.result_processor` method.
.. note::
User-defined subclasses of :class:`_types.TypeDecorator` should
**not** implement this method, and should instead implement
:meth:`_types.TypeDecorator.process_result_value` so that the
"inner" processing provided by the implementing type is maintained.
:param dialect: Dialect instance in use.
:param coltype: A SQLAlchemy data type
"""
if self._has_result_processor:
process_value = self.process_result_value
impl_processor = self.impl_instance.result_processor(
dialect, coltype
)
if impl_processor:
fixed_process_value = process_value
fixed_impl_processor = impl_processor
def process(value: Any) -> Optional[_T]:
return fixed_process_value(
fixed_impl_processor(value), dialect
)
else:
fixed_process_value = process_value
def process(value: Any) -> Optional[_T]:
return fixed_process_value(value, dialect)
return process
else:
return self.impl_instance.result_processor(dialect, coltype)
@util.memoized_property
def _has_bind_expression(self) -> bool:
return (
util.method_is_overridden(self, TypeDecorator.bind_expression)
or self.impl_instance._has_bind_expression
)
def bind_expression(
self, bindparam: BindParameter[_T]
) -> Optional[ColumnElement[_T]]:
"""Given a bind value (i.e. a :class:`.BindParameter` instance),
return a SQL expression which will typically wrap the given parameter.
.. note::
This method is called during the **SQL compilation** phase of a
statement, when rendering a SQL string. It is **not** necessarily
called against specific values, and should not be confused with the
:meth:`_types.TypeDecorator.process_bind_param` method, which is
the more typical method that processes the actual value passed to a
particular parameter at statement execution time.
Subclasses of :class:`_types.TypeDecorator` can override this method
to provide custom bind expression behavior for the type. This
implementation will **replace** that of the underlying implementation
type.
"""
return self.impl_instance.bind_expression(bindparam)
@util.memoized_property
def _has_column_expression(self) -> bool:
"""memoized boolean, check if column_expression is implemented.
Allows the method to be skipped for the vast majority of expression
types that don't use this feature.
"""
return (
util.method_is_overridden(self, TypeDecorator.column_expression)
or self.impl_instance._has_column_expression
)
def column_expression(
self, column: ColumnElement[_T]
) -> Optional[ColumnElement[_T]]:
"""Given a SELECT column expression, return a wrapping SQL expression.
.. note::
This method is called during the **SQL compilation** phase of a
statement, when rendering a SQL string. It is **not** called
against specific values, and should not be confused with the
:meth:`_types.TypeDecorator.process_result_value` method, which is
the more typical method that processes the actual value returned
in a result row subsequent to statement execution time.
Subclasses of :class:`_types.TypeDecorator` can override this method
to provide custom column expression behavior for the type. This
implementation will **replace** that of the underlying implementation
type.
See the description of :meth:`_types.TypeEngine.column_expression`
for a complete description of the method's use.
"""
return self.impl_instance.column_expression(column)
def coerce_compared_value(
self, op: Optional[OperatorType], value: Any
) -> Any:
"""Suggest a type for a 'coerced' Python value in an expression.
By default, returns self. This method is called by
the expression system when an object using this type is
on the left or right side of an expression against a plain Python
object which does not yet have a SQLAlchemy type assigned::
expr = table.c.somecolumn + 35
Where above, if ``somecolumn`` uses this type, this method will
be called with the value ``operator.add``
and ``35``. The return value is whatever SQLAlchemy type should
be used for ``35`` for this particular operation.
"""
return self
def copy(self, **kw: Any) -> Self:
"""Produce a copy of this :class:`.TypeDecorator` instance.
This is a shallow copy and is provided to fulfill part of
the :class:`.TypeEngine` contract. It usually does not
need to be overridden unless the user-defined :class:`.TypeDecorator`
has local state that should be deep-copied.
"""
instance = self.__class__.__new__(self.__class__)
instance.__dict__.update(self.__dict__)
return instance
def get_dbapi_type(self, dbapi: ModuleType) -> Optional[Any]:
"""Return the DBAPI type object represented by this
:class:`.TypeDecorator`.
By default this calls upon :meth:`.TypeEngine.get_dbapi_type` of the
underlying "impl".
"""
return self.impl_instance.get_dbapi_type(dbapi)
def compare_values(self, x: Any, y: Any) -> bool:
"""Given two values, compare them for equality.
By default this calls upon :meth:`.TypeEngine.compare_values`
of the underlying "impl", which in turn usually
uses the Python equals operator ``==``.
This function is used by the ORM to compare
an original-loaded value with an intercepted
"changed" value, to determine if a net change
has occurred.
"""
return self.impl_instance.compare_values(x, y)
# mypy property bug
@property
def sort_key_function(self) -> Optional[Callable[[Any], Any]]: # type: ignore # noqa: E501
return self.impl_instance.sort_key_function
def __repr__(self) -> str:
return util.generic_repr(self, to_inspect=self.impl_instance)
class Variant(TypeDecorator[_T]):
"""deprecated. symbol is present for backwards-compatibility with
workaround recipes, however this actual type should not be used.
"""
def __init__(self, *arg: Any, **kw: Any):
raise NotImplementedError(
"Variant is no longer used in SQLAlchemy; this is a "
"placeholder symbol for backwards compatibility."
)
@overload
def to_instance(
typeobj: Union[Type[_TE], _TE], *arg: Any, **kw: Any
) -> _TE: ...
@overload
def to_instance(typeobj: None, *arg: Any, **kw: Any) -> TypeEngine[None]: ...
def to_instance(
typeobj: Union[Type[_TE], _TE, None], *arg: Any, **kw: Any
) -> Union[_TE, TypeEngine[None]]:
if typeobj is None:
return NULLTYPE
if callable(typeobj):
return typeobj(*arg, **kw)
else:
return typeobj
def adapt_type(
typeobj: TypeEngine[Any],
colspecs: Mapping[Type[Any], Type[TypeEngine[Any]]],
) -> TypeEngine[Any]:
if isinstance(typeobj, type):
typeobj = typeobj()
for t in typeobj.__class__.__mro__[0:-1]:
try:
impltype = colspecs[t]
break
except KeyError:
pass
else:
# couldn't adapt - so just return the type itself
# (it may be a user-defined type)
return typeobj
# if we adapted the given generic type to a database-specific type,
# but it turns out the originally given "generic" type
# is actually a subclass of our resulting type, then we were already
# given a more specific type than that required; so use that.
if issubclass(typeobj.__class__, impltype):
return typeobj
return typeobj.adapt(impltype)