What’s New in SQLAlchemy 0.8?

    This document describes changes between SQLAlchemy version 0.7, undergoing maintenance releases as of October, 2012, and SQLAlchemy version 0.8, which is expected for release in early 2013.

    Document date: October 25, 2012 Updated: March 9, 2013

    This guide introduces what’s new in SQLAlchemy version 0.8, and also documents changes which affect users migrating their applications from the 0.7 series of SQLAlchemy to 0.8.

    SQLAlchemy releases are closing in on 1.0, and each new version since 0.5 features fewer major usage changes. Most applications that are settled into modern 0.7 patterns should be movable to 0.8 with no changes. Applications that use 0.6 and even 0.5 patterns should be directly migratable to 0.8 as well, though larger applications may want to test with each interim version.

    Platform Support

    SQLAlchemy 0.8 will target Python 2.5 and forward; compatibility for Python 2.4 is being dropped.

    The internals will be able to make usage of Python ternaries (that is, x if y else z) which will improve things versus the usage of y and x or z, which naturally has been the source of some bugs, as well as context managers (that is, with:) and perhaps in some cases try:/except:/else: blocks which will help with code readability.

    SQLAlchemy will eventually drop 2.5 support as well - when 2.6 is reached as the baseline, SQLAlchemy will move to use 2.6/3.3 in-place compatibility, removing the usage of the 2to3 tool and maintaining a source base that works with Python 2 and 3 at the same time.

    Rewritten mechanics

    0.8 features a much improved and capable system regarding how relationship() determines how to join between two entities. The new system includes these features:

    • The primaryjoin argument is no longer needed when constructing a against a class that has multiple foreign key paths to the target. Only the foreign_keys argument is needed to specify those columns which should be included:

      ``` class Parent(Base):

    1. class Child(Base):
    2. __tablename__ = "child"
    3. id = Column(Integer, primary_key=True)
    4. ```
    • relationships against self-referential, composite foreign keys where a column points to itself are now supported. The canonical case is as follows:

      1. class Folder(Base):
      2. __tablename__ = "folder"
      3. __table_args__ = (
      4. ForeignKeyConstraint(
      5. ["account_id", "parent_id"], ["folder.account_id", "folder.folder_id"]
      6. ),
      7. )
      8. account_id = Column(Integer, primary_key=True)
      9. folder_id = Column(Integer, primary_key=True)
      10. parent_id = Column(Integer)
      11. name = Column(String)
      12. parent_folder = relationship(
      13. "Folder", backref="child_folders", remote_side=[account_id, folder_id]
      14. )

      Above, the Folder refers to its parent Folder joining from account_id to itself, and parent_id to folder_id. When SQLAlchemy constructs an auto- join, no longer can it assume all columns on the “remote” side are aliased, and all columns on the “local” side are not - the account_id column is on both sides. So the internal relationship mechanics were totally rewritten to support an entirely different system whereby two copies of account_id are generated, each containing different annotations to determine their role within the statement. Note the join condition within a basic eager load:

      1. SELECT
      2. folder.account_id AS folder_account_id,
      3. folder.folder_id AS folder_folder_id,
      4. folder.parent_id AS folder_parent_id,
      5. folder.name AS folder_name,
      6. folder_1.account_id AS folder_1_account_id,
      7. folder_1.folder_id AS folder_1_folder_id,
      8. folder_1.parent_id AS folder_1_parent_id,
      9. folder_1.name AS folder_1_name
      10. FROM folder
      11. LEFT OUTER JOIN folder AS folder_1
      12. ON
      13. folder_1.account_id = folder.account_id
      14. AND folder.folder_id = folder_1.parent_id
      15. WHERE folder.folder_id = ? AND folder.account_id = ?
    • Previously difficult custom join conditions, like those involving functions and/or CASTing of types, will now function as expected in most cases:

      1. class HostEntry(Base):
      2. __tablename__ = "host_entry"
      3. id = Column(Integer, primary_key=True)
      4. ip_address = Column(INET)
      5. content = Column(String(50))
      6. # relationship() using explicit foreign_keys, remote_side
      7. parent_host = relationship(
      8. "HostEntry",
      9. primaryjoin=ip_address == cast(content, INET),
      10. foreign_keys=content,
      11. remote_side=ip_address,
      12. )

      The new relationship() mechanics make use of a SQLAlchemy concept known as . These annotations are also available to application code explicitly via the foreign() and functions, either as a means to improve readability for advanced configurations or to directly inject an exact configuration, bypassing the usual join-inspection heuristics:

      ``` from sqlalchemy.orm import foreign, remote

    1. class HostEntry(Base):
    2. __tablename__ = "host_entry"
    3. id = Column(Integer, primary_key=True)
    4. ip_address = Column(INET)
    5. content = Column(String(50))
    6. # relationship() using explicit foreign() and remote() annotations
    7. # in lieu of separate arguments
    8. parent_host = relationship(
    9. "HostEntry",
    10. primaryjoin=remote(ip_address) == cast(foreign(content), INET),
    11. )
    12. ```

    See also

    Configuring how Relationship Joins - a newly revised section on detailing the latest techniques for customizing related attributes and collection access.

    #1401

    New Class/Object Inspection System

    Lots of SQLAlchemy users are writing systems that require the ability to inspect the attributes of a mapped class, including being able to get at the primary key columns, object relationships, plain attributes, and so forth, typically for the purpose of building data-marshalling systems, like JSON/XML conversion schemes and of course form libraries galore.

    Originally, the and Column model were the original inspection points, which have a well-documented system. While SQLAlchemy ORM models are also fully introspectable, this has never been a fully stable and supported feature, and users tended to not have a clear idea how to get at this information.

    0.8 now provides a consistent, stable and fully documented API for this purpose, including an inspection system which works on mapped classes, instances, attributes, and other Core and ORM constructs. The entrypoint to this system is the core-level function. In most cases, the object being inspected is one already part of SQLAlchemy’s system, such as Mapper, , Inspector. In some cases, new objects have been added with the job of providing the inspection API in certain contexts, such as and AttributeState.

    A walkthrough of some key capabilities follows:

    1. >>> class User(Base):
    2. ... __tablename__ = "user"
    3. ... id = Column(Integer, primary_key=True)
    4. ... name = Column(String)
    5. ... name_syn = synonym(name)
    6. ... addresses = relationship("Address")
    7. >>> # universal entry point is inspect()
    8. >>> b = inspect(User)
    9. >>> # b in this case is the Mapper
    10. >>> b
    11. <Mapper at 0x101521950; User>
    12. >>> # Column namespace
    13. >>> b.columns.id
    14. Column('id', Integer(), table=<user>, primary_key=True, nullable=False)
    15. >>> # mapper's perspective of the primary key
    16. >>> b.primary_key
    17. (Column('id', Integer(), table=<user>, primary_key=True, nullable=False),)
    18. >>> # MapperProperties available from .attrs
    19. >>> b.attrs.keys()
    20. ['name_syn', 'addresses', 'id', 'name']
    21. >>> # .column_attrs, .relationships, etc. filter this collection
    22. >>> b.column_attrs.keys()
    23. ['id', 'name']
    24. >>> list(b.relationships)
    25. [<sqlalchemy.orm.properties.RelationshipProperty object at 0x1015212d0>]
    26. >>> # they are also namespaces
    27. >>> b.column_attrs.id
    28. <sqlalchemy.orm.properties.ColumnProperty object at 0x101525090>
    29. >>> b.relationships.addresses
    30. <sqlalchemy.orm.properties.RelationshipProperty object at 0x1015212d0>
    31. >>> # point inspect() at a mapped, class level attribute,
    32. >>> # returns the attribute itself
    33. >>> b = inspect(User.addresses)
    34. >>> b
    35. <sqlalchemy.orm.attributes.InstrumentedAttribute object at 0x101521fd0>
    36. >>> # From here we can get the mapper:
    37. >>> b.mapper
    38. <Mapper at 0x101525810; Address>
    39. >>> # the parent inspector, in this case a mapper
    40. >>> b.parent
    41. <Mapper at 0x101521950; User>
    42. >>> # an expression
    43. >>> print(b.expression)
    44. "user".id = address.user_id
    45. >>> # inspect works on instances
    46. >>> u1 = User(id=3, name="x")
    47. >>> b = inspect(u1)
    48. >>> # it returns the InstanceState
    49. >>> b
    50. <sqlalchemy.orm.state.InstanceState object at 0x10152bed0>
    51. >>> # similar attrs accessor refers to the
    52. >>> b.attrs.keys()
    53. ['id', 'name_syn', 'addresses', 'name']
    54. >>> # attribute interface - from attrs, you get a state object
    55. >>> b.attrs.id
    56. <sqlalchemy.orm.state.AttributeState object at 0x10152bf90>
    57. >>> # this object can give you, current value...
    58. >>> b.attrs.id.value
    59. 3
    60. >>> # ... current history
    61. >>> b.attrs.id.history
    62. History(added=[3], unchanged=(), deleted=())
    63. >>> # InstanceState can also provide session state information
    64. >>> # lets assume the object is persistent
    65. >>> s = Session()
    66. >>> s.add(u1)
    67. >>> s.commit()
    68. >>> # now we can get primary key identity, always
    69. >>> # works in query.get()
    70. >>> b.identity
    71. (3,)
    72. >>> # the mapper level key
    73. >>> b.identity_key
    74. (<class '__main__.User'>, (3,))
    75. >>> # state within the session
    76. >>> b.persistent, b.transient, b.deleted, b.detached
    77. (True, False, False, False)
    78. >>> # owning session
    79. >>> b.session
    80. <sqlalchemy.orm.session.Session object at 0x101701150>

    See also

    #2208

    New with_polymorphic() feature, can be used anywhere

    The Query.with_polymorphic() method allows the user to specify which tables should be present when querying against a joined-table entity. Unfortunately the method is awkward and only applies to the first entity in the list, and otherwise has awkward behaviors both in usage as well as within the internals. A new enhancement to the aliased() construct has been added called which allows any entity to be “aliased” into a “polymorphic” version of itself, freely usable anywhere:

    1. from sqlalchemy.orm import with_polymorphic
    2. palias = with_polymorphic(Person, [Engineer, Manager])
    3. session.query(Company).join(palias, Company.employees).filter(
    4. or_(Engineer.language == "java", Manager.hair == "pointy")
    5. )

    See also

    Using with_polymorphic() - newly updated documentation for polymorphic loading control.

    of_type() works with alias(), with_polymorphic(), any(), has(), joinedload(), subqueryload(), contains_eager()

    The method is used to specify a specific subtype to use when constructing SQL expressions along a relationship() that has a mapping as its target. This method can now be used to target any number of target subtypes, by combining it with the new with_polymorphic() function:

    1. # use eager loading in conjunction with with_polymorphic targets
    2. Job_P = with_polymorphic(Job, [SubJob, ExtraJob], aliased=True)
    3. q = (
    4. s.query(DataContainer)
    5. .join(DataContainer.jobs.of_type(Job_P))
    6. .options(contains_eager(DataContainer.jobs.of_type(Job_P)))
    7. )

    The method now works equally well in most places a regular relationship attribute is accepted, including with loader functions like , subqueryload(), , and comparison methods like PropComparator.any() and :

    1. Job_P = with_polymorphic(Job, [SubJob, ExtraJob], aliased=True)
    2. q = (
    3. s.query(DataContainer)
    4. .join(DataContainer.jobs.of_type(Job_P))
    5. .options(contains_eager(DataContainer.jobs.of_type(Job_P)))
    6. )
    7. # pass subclasses to eager loads (implicitly applies with_polymorphic)
    8. q = s.query(ParentThing).options(
    9. joinedload_all(ParentThing.container, DataContainer.jobs.of_type(SubJob))
    10. )
    11. # control self-referential aliasing with any()/has()
    12. Job_A = aliased(Job)
    13. q = (
    14. s.query(Job)
    15. .join(DataContainer.jobs)
    16. .filter(
    17. DataContainer.jobs.of_type(Job_A).any(
    18. and_(Job_A.id < Job.id, Job_A.type == "fred")
    19. )
    20. )
    21. )

    See also

    Joining to specific sub-types or with_polymorphic() entities

    #1106

    Events Can Be Applied to Unmapped Superclasses

    Mapper and instance events can now be associated with an unmapped superclass, where those events will be propagated to subclasses as those subclasses are mapped. The propagate=True flag should be used. This feature allows events to be associated with a declarative base class:

    1. from sqlalchemy.ext.declarative import declarative_base
    2. Base = declarative_base()
    3. @event.listens_for("load", Base, propagate=True)
    4. def on_load(target, context):
    5. print("New instance loaded:", target)
    6. # on_load() will be applied to SomeClass
    7. class SomeClass(Base):
    8. __tablename__ = "sometable"
    9. # ...

    #2585

    Declarative Distinguishes Between Modules/Packages

    A key feature of Declarative is the ability to refer to other mapped classes using their string name. The registry of class names is now sensitive to the owning module and package of a given class. The classes can be referred to via dotted name in expressions:

    1. class Snack(Base):
    2. # ...
    3. peanuts = relationship(
    4. "nuts.Peanut", primaryjoin="nuts.Peanut.snack_id == Snack.id"
    5. )

    The resolution allows that any full or partial disambiguating package name can be used. If the path to a particular class is still ambiguous, an error is raised.

    #2338

    New DeferredReflection Feature in Declarative

    The “deferred reflection” example has been moved to a supported feature within Declarative. This feature allows the construction of declarative mapped classes with only placeholder Table metadata, until a prepare() step is called, given an Engine with which to reflect fully all tables and establish actual mappings. The system supports overriding of columns, single and joined inheritance, as well as distinct bases-per-engine. A full declarative configuration can now be created against an existing table that is assembled upon engine creation time in one step:

    1. class ReflectedOne(DeferredReflection, Base):
    2. __abstract__ = True
    3. class ReflectedTwo(DeferredReflection, Base):
    4. __abstract__ = True
    5. class MyClass(ReflectedOne):
    6. __tablename__ = "mytable"
    7. class MyOtherClass(ReflectedOne):
    8. __tablename__ = "myothertable"
    9. class YetAnotherClass(ReflectedTwo):
    10. __tablename__ = "yetanothertable"
    11. ReflectedOne.prepare(engine_one)
    12. ReflectedTwo.prepare(engine_two)

    See also

    DeferredReflection

    ORM Classes Now Accepted by Core Constructs

    While the SQL expressions used with , such as User.id == 5, have always been compatible for use with core constructs such as select(), the mapped class itself would not be recognized when passed to , Select.select_from(), or . A new SQL registration system allows a mapped class to be accepted as a FROM clause within the core:

    1. from sqlalchemy import select
    2. stmt = select([User]).where(User.id == 5)

    Above, the mapped User class will expand into the Table to which User is mapped.

    Query.update() supports UPDATE..FROM

    The new UPDATE..FROM mechanics work in query.update(). Below, we emit an UPDATE against SomeEntity, adding a FROM clause (or equivalent, depending on backend) against SomeOtherEntity:

    1. query(SomeEntity).filter(SomeEntity.id == SomeOtherEntity.id).filter(
    2. SomeOtherEntity.foo == "bar"
    3. ).update({"data": "x"})

    In particular, updates to joined-inheritance entities are supported, provided the target of the UPDATE is local to the table being filtered on, or if the parent and child tables are mixed, they are joined explicitly in the query. Below, given Engineer as a joined subclass of Person:

    1. query(Engineer).filter(Person.id == Engineer.id).filter(
    2. Person.name == "dilbert"
    3. ).update({"engineer_data": "java"})

    would produce:

    1. UPDATE engineer SET engineer_data='java' FROM person
    2. WHERE person.id=engineer.id AND person.name='dilbert'

    rollback() will only roll back “dirty” objects from a begin_nested()

    A behavioral change that should improve efficiency for those users using SAVEPOINT via Session.begin_nested() - upon rollback(), only those objects that were made dirty since the last flush will be expired, the rest of the Session remains intact. This because a ROLLBACK to a SAVEPOINT does not terminate the containing transaction’s isolation, so no expiry is needed except for those changes that were not flushed in the current transaction.

    Caching Example now uses dogpile.cache

    The caching example now uses . Dogpile.cache is a rewrite of the caching portion of Beaker, featuring vastly simpler and faster operation, as well as support for distributed locking.

    Note that the SQLAlchemy APIs used by the Dogpile example as well as the previous Beaker example have changed slightly, in particular this change is needed as illustrated in the Beaker example:

    See also

    Dogpile Caching

    New Core Features

    Fully extensible, type-level operator support in Core

    The Core has to date never had any system of adding support for new SQL operators to Column and other expression constructs, other than the ColumnOperators.op() method which is “just enough” to make things work. There has also never been any system in place for Core which allows the behavior of existing operators to be overridden. Up until now, the only way operators could be flexibly redefined was in the ORM layer, using given a comparator_factory argument. Third party libraries like GeoAlchemy therefore were forced to be ORM-centric and rely upon an array of hacks to apply new operations as well as to get them to propagate correctly.

    For example, to add logarithm support to Numeric types:

    1. from sqlalchemy.types import Numeric
    2. from sqlalchemy.sql import func
    3. class CustomNumeric(Numeric):
    4. class comparator_factory(Numeric.Comparator):
    5. def log(self, other):
    6. return func.log(self.expr, other)

    The new type is usable like any other type:

    1. data = Table(
    2. "data",
    3. metadata,
    4. Column("id", Integer, primary_key=True),
    5. Column("x", CustomNumeric(10, 5)),
    6. Column("y", CustomNumeric(10, 5)),
    7. )
    8. stmt = select([data.c.x.log(data.c.y)]).where(data.c.x.log(2) < value)
    9. print(conn.execute(stmt).fetchall())

    New features which have come from this immediately include support for PostgreSQL’s HSTORE type, as well as new operations associated with PostgreSQL’s ARRAY type. It also paves the way for existing types to acquire lots more operators that are specific to those types, such as more string, integer and date operators.

    See also

    HSTORE

    The method now supports a list of dictionaries, which will render a multi-VALUES statement such as VALUES (<row1>), (<row2>), .... This is only relevant to backends which support this syntax, including PostgreSQL, SQLite, and MySQL. It is not the same thing as the usual executemany() style of INSERT which remains unchanged:

    1. users.insert().values(
    2. [
    3. {"name": "some name"},
    4. {"name": "some other name"},
    5. {"name": "yet another name"},
    6. ]
    7. )

    See also

    Insert.values()

    Type Expressions

    SQL expressions can now be associated with types. Historically, has always allowed Python-side functions which receive both bound parameters as well as result row values, passing them through a Python side conversion function on the way to/back from the database. The new feature allows similar functionality, except on the database side:

    1. from sqlalchemy.types import String
    2. from sqlalchemy import func, Table, Column, MetaData
    3. class LowerString(String):
    4. def bind_expression(self, bindvalue):
    5. return func.lower(bindvalue)
    6. def column_expression(self, col):
    7. return func.lower(col)
    8. metadata = MetaData()
    9. test_table = Table("test_table", metadata, Column("data", LowerString))

    Above, the LowerString type defines a SQL expression that will be emitted whenever the test_table.c.data column is rendered in the columns clause of a SELECT statement:

    1. >>> print(select([test_table]).where(test_table.c.data == "HI"))
    2. SELECT lower(test_table.data) AS data
    3. FROM test_table
    4. WHERE test_table.data = lower(:data_1)

    This feature is also used heavily by the new release of GeoAlchemy, to embed PostGIS expressions inline in SQL based on type rules.

    See also

    Applying SQL-level Bind/Result Processing

    Core Inspection System

    The function introduced in New Class/Object Inspection System also applies to the core. Applied to an it produces an Inspector object:

    1. from sqlalchemy import inspect
    2. from sqlalchemy import create_engine
    3. engine = create_engine("postgresql://scott:tiger@localhost/test")
    4. insp = inspect(engine)
    5. print(insp.get_table_names())

    It can also be applied to any , which returns the ClauseElement itself, such as , Column, , etc. This allows it to work fluently between Core and ORM constructs.

    New Method

    select() now has a method which specifies “correlate on all FROM clauses except those specified”. It can be used for mapping scenarios where a related subquery should correlate normally, except against a particular target selectable:

    1. class SnortEvent(Base):
    2. __tablename__ = "event"
    3. id = Column(Integer, primary_key=True)
    4. signature = Column(Integer, ForeignKey("signature.id"))
    5. signatures = relationship("Signature", lazy=False)
    6. class Signature(Base):
    7. __tablename__ = "signature"
    8. id = Column(Integer, primary_key=True)
    9. sig_count = column_property(
    10. select([func.count("*")])
    11. .where(SnortEvent.signature == id)
    12. .correlate_except(SnortEvent)
    13. )

    See also

    Select.correlate_except()

    PostgreSQL HSTORE type

    Support for PostgreSQL’s HSTORE type is now available as HSTORE. This type makes great usage of the new operator system to provide a full range of operators for HSTORE types, including index access, concatenation, and containment methods such as comparator_factory.has_key(), comparator_factory.has_any(), and comparator_factory.matrix():

    1. from sqlalchemy.dialects.postgresql import HSTORE
    2. data = Table(
    3. "data_table",
    4. metadata,
    5. Column("id", Integer, primary_key=True),
    6. Column("hstore_data", HSTORE),
    7. )
    8. engine.execute(select([data.c.hstore_data["some_key"]])).scalar()
    9. engine.execute(select([data.c.hstore_data.matrix()])).scalar()

    See also

    hstore

    Enhanced PostgreSQL ARRAY type

    The type will accept an optional “dimension” argument, pinning it to a fixed number of dimensions and greatly improving efficiency when retrieving results:

    1. # old way, still works since PG supports N-dimensions per row:
    2. Column("my_array", postgresql.ARRAY(Integer))
    3. # new way, will render ARRAY with correct number of [] in DDL,
    4. # will process binds and results more efficiently as we don't need
    5. # to guess how many levels deep to go
    6. Column("my_array", postgresql.ARRAY(Integer, dimensions=2))

    The type also introduces new operators, using the new type-specific operator framework. New operations include indexed access:

    1. result = conn.execute(select([mytable.c.arraycol[2]]))

    slice access in SELECT:

    1. result = conn.execute(select([mytable.c.arraycol[2:4]]))

    slice updates in UPDATE:

    1. conn.execute(mytable.update().values({mytable.c.arraycol[2:3]: [7, 8]}))

    freestanding array literals:

    1. >>> from sqlalchemy.dialects import postgresql
    2. >>> conn.scalar(select([postgresql.array([1, 2]) + postgresql.array([3, 4, 5])]))
    3. [1, 2, 3, 4, 5]

    array concatenation, where below, the right side [4, 5, 6] is coerced into an array literal:

    1. select([mytable.c.arraycol + [4, 5, 6]])

    See also

    ARRAY

    #2441

    New, configurable DATE, TIME types for SQLite

    SQLite has no built-in DATE, TIME, or DATETIME types, and instead provides some support for storage of date and time values either as strings or integers. The date and time types for SQLite are enhanced in 0.8 to be much more configurable as to the specific format, including that the “microseconds” portion is optional, as well as pretty much everything else.

    1. Column("sometimestamp", sqlite.DATETIME(truncate_microseconds=True))
    2. Column(
    3. "sometimestamp",
    4. storage_format=(
    5. "%(year)04d%(month)02d%(day)02d"
    6. "%(hour)02d%(minute)02d%(second)02d%(microsecond)06d"
    7. ),
    8. regexp="(\d{4})(\d{2})(\d{2})(\d{2})(\d{2})(\d{2})(\d{6})",
    9. ),
    10. )
    11. Column(
    12. "somedate",
    13. sqlite.DATE(
    14. storage_format="%(month)02d/%(day)02d/%(year)04d",
    15. regexp="(?P<month>\d+)/(?P<day>\d+)/(?P<year>\d+)",
    16. ),
    17. )

    Huge thanks to Nate Dub for the sprinting on this at Pycon 2012.

    See also

    DATETIME

    TIME

    “COLLATE” supported across all dialects; in particular MySQL, PostgreSQL, SQLite

    The “collate” keyword, long accepted by the MySQL dialect, is now established on all types and will render on any backend, including when features such as MetaData.create_all() and is used:

    1. >>> stmt = select([cast(sometable.c.somechar, String(20, collation="utf8"))])
    2. >>> print(stmt)
    3. SELECT CAST(sometable.somechar AS VARCHAR(20) COLLATE "utf8") AS anon_1
    4. FROM sometable

    See also

    String

    “Prefixes” now supported for , delete()

    Geared towards MySQL, a “prefix” can be rendered within any of these constructs. E.g.:

    The method is new in addition to those which already existed on , select() and .

    See also

    Update.prefix_with()

    Delete.prefix_with()

    Insert.prefix_with()

    Select.prefix_with()

    #2431

    The consideration of a “pending” object as an “orphan” has been made more aggressive

    This is a late add to the 0.8 series, however it is hoped that the new behavior is generally more consistent and intuitive in a wider variety of situations. The ORM has since at least version 0.4 included behavior such that an object that’s “pending”, meaning that it’s associated with a but hasn’t been inserted into the database yet, is automatically expunged from the Session when it becomes an “orphan”, which means it has been de-associated with a parent object that refers to it with delete-orphan cascade on the configured . This behavior is intended to approximately mirror the behavior of a persistent (that is, already inserted) object, where the ORM will emit a DELETE for such objects that become orphans based on the interception of detachment events.

    The rationale for the older behavior dates back at least to version 0.4, and was basically a defensive decision to try to alleviate confusion when an object was still being constructed for INSERT. But the reality is that the object is re-associated with the Session as soon as it is attached to any new parent in any case.

    It’s still possible to flush an object that is not associated with all of its required parents, if the object was either not associated with those parents in the first place, or if it was expunged, but then re-associated with a via a subsequent attachment event but still not fully associated. In this situation, it is expected that the database would emit an integrity error, as there are likely NOT NULL foreign key columns that are unpopulated. The ORM makes the decision to let these INSERT attempts occur, based on the judgment that an object that is only partially associated with its required parents but has been actively associated with some of them, is more often than not a user error, rather than an intentional omission which should be silently skipped - silently skipping the INSERT here would make user errors of this nature very hard to debug.

    The old behavior, for applications that might have been relying upon it, can be re-enabled for any Mapper by specifying the flag legacy_is_orphan as a mapper option.

    The new behavior allows the following test case to work:

    1. from sqlalchemy import Column, Integer, String, ForeignKey
    2. from sqlalchemy.orm import relationship, backref
    3. from sqlalchemy.ext.declarative import declarative_base
    4. Base = declarative_base()
    5. class User(Base):
    6. __tablename__ = "user"
    7. id = Column(Integer, primary_key=True)
    8. name = Column(String(64))
    9. class UserKeyword(Base):
    10. __tablename__ = "user_keyword"
    11. user_id = Column(Integer, ForeignKey("user.id"), primary_key=True)
    12. keyword_id = Column(Integer, ForeignKey("keyword.id"), primary_key=True)
    13. user = relationship(
    14. User, backref=backref("user_keywords", cascade="all, delete-orphan")
    15. )
    16. keyword = relationship(
    17. "Keyword", backref=backref("user_keywords", cascade="all, delete-orphan")
    18. )
    19. # uncomment this to enable the old behavior
    20. # __mapper_args__ = {"legacy_is_orphan": True}
    21. class Keyword(Base):
    22. __tablename__ = "keyword"
    23. id = Column(Integer, primary_key=True)
    24. keyword = Column("keyword", String(64))
    25. from sqlalchemy import create_engine
    26. from sqlalchemy.orm import Session
    27. # note we're using PostgreSQL to ensure that referential integrity
    28. # is enforced, for demonstration purposes.
    29. e = create_engine("postgresql://scott:tiger@localhost/test", echo=True)
    30. Base.metadata.drop_all(e)
    31. Base.metadata.create_all(e)
    32. session = Session(e)
    33. u1 = User(name="u1")
    34. k1 = Keyword(keyword="k1")
    35. session.add_all([u1, k1])
    36. uk1 = UserKeyword(keyword=k1, user=u1)
    37. # previously, if session.flush() were called here,
    38. # this operation would succeed, but if session.flush()
    39. # were not called here, the operation fails with an
    40. # integrity error.
    41. # session.flush()
    42. del u1.user_keywords[0]
    43. session.commit()

    The after_attach event fires after the item is associated with the Session instead of before; before_attach added

    Event handlers which use after_attach can now assume the given instance is associated with the given session:

    1. @event.listens_for(Session, "after_attach")
    2. def after_attach(session, instance):
    3. assert instance in session

    Some use cases require that it work this way. However, other use cases require that the item is not yet part of the session, such as when a query, intended to load some state required for an instance, emits autoflush first and would otherwise prematurely flush the target object. Those use cases should use the new “before_attach” event:

    1. @event.listens_for(Session, "before_attach")
    2. def before_attach(session, instance):
    3. instance.some_necessary_attribute = (
    4. session.query(Widget).filter_by(instance.widget_name).first()
    5. )

    Query now auto-correlates like a select() does

    Previously it was necessary to call in order to have a column- or WHERE-subquery correlate to the parent:

    1. subq = (
    2. session.query(Entity.value)
    3. .filter(Entity.id == Parent.entity_id)
    4. .correlate(Parent)
    5. .as_scalar()
    6. )
    7. session.query(Parent).filter(subq == "some value")

    This was the opposite behavior of a plain select() construct which would assume auto-correlation by default. The above statement in 0.8 will correlate automatically:

    1. subq = session.query(Entity.value).filter(Entity.id == Parent.entity_id).as_scalar()
    2. session.query(Parent).filter(subq == "some value")

    like in select(), correlation can be disabled by calling query.correlate(None) or manually set by passing an entity, query.correlate(someentity).

    #2179

    Correlation is now always context-specific

    To allow a wider variety of correlation scenarios, the behavior of Select.correlate() and has changed slightly such that the SELECT statement will omit the “correlated” target from the FROM clause only if the statement is actually used in that context. Additionally, it’s no longer possible for a SELECT statement that’s placed as a FROM in an enclosing SELECT statement to “correlate” (i.e. omit) a FROM clause.

    This change only makes things better as far as rendering SQL, in that it’s no longer possible to render illegal SQL where there are insufficient FROM objects relative to what’s being selected:

    1. from sqlalchemy.sql import table, column, select
    2. t1 = table("t1", column("x"))
    3. t2 = table("t2", column("y"))
    4. s = select([t1, t2]).correlate(t1)
    5. print(s)

    Prior to this change, the above would return:

    1. SELECT t1.x, t2.y FROM t2

    which is invalid SQL as “t1” is not referred to in any FROM clause.

    Now, in the absence of an enclosing SELECT, it returns:

    1. SELECT t1.x, t2.y FROM t1, t2

    Within a SELECT, the correlation takes effect as expected:

    1. s2 = select([t1, t2]).where(t1.c.x == t2.c.y).where(t1.c.x == s)
    2. print(s2)
    1. SELECT t1.x, t2.y FROM t1, t2
    2. WHERE t1.x = t2.y AND t1.x =
    3. (SELECT t1.x, t2.y FROM t2)

    This change is not expected to impact any existing applications, as the correlation behavior remains identical for properly constructed expressions. Only an application that relies, most likely within a testing scenario, on the invalid string output of a correlated SELECT used in a non-correlating context would see any change.

    #2668

    The methods MetaData.create_all() and will now accept a list of Table objects that is empty, and will not emit any CREATE or DROP statements. Previously, an empty list was interpreted the same as passing None for a collection, and CREATE/DROP would be emitted for all items unconditionally.

    This is a bug fix but some applications may have been relying upon the previous behavior.

    Repaired the Event Targeting of

    The InstrumentationEvents series of event targets have documented that the events will only be fired off according to the actual class passed as a target. Through 0.7, this wasn’t the case, and any event listener applied to would be invoked for all classes mapped. In 0.8, additional logic has been added so that the events will only invoke for those classes sent in. The propagate flag here is set to True by default as class instrumentation events are typically used to intercept classes that aren’t yet created.

    #2590

    No more magic coercion of “=” to IN when comparing to subquery in MS-SQL

    We found a very old behavior in the MSSQL dialect which would attempt to rescue users from themselves when doing something like this:

    1. scalar_subq = select([someothertable.c.id]).where(someothertable.c.data == "foo")
    2. select([sometable]).where(sometable.c.id == scalar_subq)

    SQL Server doesn’t allow an equality comparison to a scalar SELECT, that is, “x = (SELECT something)”. The MSSQL dialect would convert this to an IN. The same thing would happen however upon a comparison like “(SELECT something) = x”, and overall this level of guessing is outside of SQLAlchemy’s usual scope so the behavior is removed.

    #2277

    Fixed the behavior of Session.is_modified()

    The method accepts an argument passive which basically should not be necessary, the argument in all cases should be the value True - when left at its default of False it would have the effect of hitting the database, and often triggering autoflush which would itself change the results. In 0.8 the passive argument will have no effect, and unloaded attributes will never be checked for history since by definition there can be no pending state change on an unloaded attribute.

    See also

    Session.is_modified()

    is honored in the Select.c attribute of with Select.apply_labels()

    Users of the expression system know that Select.apply_labels() prepends the table name to each column name, affecting the names that are available from Select.c:

    1. s = select([table1]).apply_labels()
    2. s.c.table1_col1
    3. s.c.table1_col2

    Before 0.8, if the had a different Column.key, this key would be ignored, inconsistently versus when Select.apply_labels() were not used:

    1. # before 0.8
    2. table1 = Table("t1", metadata, Column("col1", Integer, key="column_one"))
    3. s = select([table1])
    4. s.c.column_one # would be accessible like this
    5. s.c.col1 # would raise AttributeError
    6. s = select([table1]).apply_labels()
    7. s.c.table1_column_one # would raise AttributeError
    8. s.c.table1_col1 # would be accessible like this

    In 0.8, is honored in both cases:

    1. # with 0.8
    2. table1 = Table("t1", metadata, Column("col1", Integer, key="column_one"))
    3. s = select([table1])
    4. s.c.column_one # works
    5. s.c.col1 # AttributeError
    6. s = select([table1]).apply_labels()
    7. s.c.table1_column_one # works
    8. s.c.table1_col1 # AttributeError

    All other behavior regarding “name” and “key” are the same, including that the rendered SQL will still use the form <tablename>_<colname> - the emphasis here was on preventing the Column.key contents from being rendered into the SELECT statement so that there are no issues with special/ non-ascii characters used in the .

    #2397

    single_parent warning is now an error

    A relationship() that is many-to-one or many-to-many and specifies “cascade=’all, delete-orphan’”, which is an awkward but nonetheless supported use case (with restrictions) will now raise an error if the relationship does not specify the single_parent=True option. Previously it would only emit a warning, but a failure would follow almost immediately within the attribute system in any case.

    Adding the inspector argument to the column_reflect event

    0.7 added a new event called column_reflect, provided so that the reflection of columns could be augmented as each one were reflected. We got this event slightly wrong in that the event gave no way to get at the current Inspector and Connection being used for the reflection, in the case that additional information from the database is needed. As this is a new event not widely used yet, we’ll be adding the inspector argument into it directly:

    1. @event.listens_for(Table, "column_reflect")
    2. def listen_for_col(inspector, table, column_info):
    3. ...

    Disabling auto-detect of collations, casing for MySQL

    The MySQL dialect does two calls, one very expensive, to load all possible collations from the database as well as information on casing, the first time an Engine connects. Neither of these collections are used for any SQLAlchemy functions, so these calls will be changed to no longer be emitted automatically. Applications that might have relied on these collections being present on engine.dialect will need to call upon _detect_collations() and _detect_casing() directly.

    “Unconsumed column names” warning becomes an exception

    Referring to a non-existent column in an insert() or update() construct will raise an error instead of a warning:

    1. t1 = table("t1", column("x"))
    2. t1.insert().values(x=5, z=5) # raises "Unconsumed column names: z"

    Inspector.get_primary_keys() is deprecated, use Inspector.get_pk_constraint

    These two methods on Inspector were redundant, where get_primary_keys() would return the same information as get_pk_constraint() minus the name of the constraint:

    Case-insensitive result row names will be disabled in most cases

    A very old behavior, the column names in RowProxy were always compared case-insensitively:

    1. >>> row = result.fetchone()
    2. >>> row["foo"] == row["FOO"] == row["Foo"]

    This was for the benefit of a few dialects which in the early days needed this, like Oracle and Firebird, but in modern usage we have more accurate ways of dealing with the case-insensitive behavior of these two platforms.

    Going forward, this behavior will be available only optionally, by passing the flag `case_sensitive=False` to `create_engine()` , but otherwise column names requested from the row must match as far as casing.

    InstrumentationManager and alternate class instrumentation is now an extension

    The sqlalchemy.orm.interfaces.InstrumentationManager class is moved to sqlalchemy.ext.instrumentation.InstrumentationManager. The “alternate instrumentation” system was built for the benefit of a very small number of installations that needed to work with existing or unusual class instrumentation systems, and generally is very seldom used. The complexity of this system has been exported to an ext. module. It remains unused until once imported, typically when a third party library imports InstrumentationManager, at which point it is injected back into sqlalchemy.orm by replacing the default InstrumentationFactory with ExtendedInstrumentationRegistry.

    Removed

    SQLSoup

    SQLSoup is a handy package that presents an alternative interface on top of the SQLAlchemy ORM. SQLSoup is now moved into its own project and documented/released separately; see .

    SQLSoup is a very simple tool that could also benefit from contributors who are interested in its style of usage.

    #2262

    The older “mutable” system within the SQLAlchemy ORM has been removed. This refers to the MutableType interface which was applied to types such as PickleType and conditionally to TypeDecorator, and since very early SQLAlchemy versions has provided a way for the ORM to detect changes in so-called “mutable” data structures such as JSON structures and pickled objects. However, the implementation was never reasonable and forced a very inefficient mode of usage on the unit-of-work which caused an expensive scan of all objects to take place during flush. In 0.7, the sqlalchemy.ext.mutable extension was introduced so that user-defined datatypes can appropriately send events to the unit of work as changes occur.

    Today, usage of MutableType is expected to be low, as warnings have been in place for some years now regarding its inefficiency.

    sqlalchemy.exceptions (has been sqlalchemy.exc for years)