{"diffoscope-json-version": 1, "source1": "/srv/reproducible-results/rbuild-debian/r-b-build.9RZY0lOO/b1/sqlalchemy_2.0.32+ds1-1_i386.changes", "source2": "/srv/reproducible-results/rbuild-debian/r-b-build.9RZY0lOO/b2/sqlalchemy_2.0.32+ds1-1_i386.changes", "unified_diff": null, "details": [{"source1": "Files", "source2": "Files", "unified_diff": "@@ -1,5 +1,5 @@\n \n- 98a02c5c49bd80e90e3e1d1c45d8c8b2 3956068 doc optional python-sqlalchemy-doc_2.0.32+ds1-1_all.deb\n+ a9d07aad0784f5e68858537e33b02d63 3956116 doc optional python-sqlalchemy-doc_2.0.32+ds1-1_all.deb\n 6be7382861298efb5a82b60405fa7f83 1748792 debug optional python3-sqlalchemy-ext-dbgsym_2.0.32+ds1-1_i386.deb\n b5c87ce5c170577275eeeed1855128e0 219664 python optional python3-sqlalchemy-ext_2.0.32+ds1-1_i386.deb\n e1c78ec120d9d481e2a5c4c579530013 1196072 python optional python3-sqlalchemy_2.0.32+ds1-1_all.deb\n"}, {"source1": "python-sqlalchemy-doc_2.0.32+ds1-1_all.deb", "source2": "python-sqlalchemy-doc_2.0.32+ds1-1_all.deb", "unified_diff": null, "details": [{"source1": "file list", "source2": "file list", "unified_diff": "@@ -1,3 +1,3 @@\n -rw-r--r-- 0 0 0 4 2024-08-23 07:52:58.000000 debian-binary\n--rw-r--r-- 0 0 0 13920 2024-08-23 07:52:58.000000 control.tar.xz\n--rw-r--r-- 0 0 0 3941956 2024-08-23 07:52:58.000000 data.tar.xz\n+-rw-r--r-- 0 0 0 13908 2024-08-23 07:52:58.000000 control.tar.xz\n+-rw-r--r-- 0 0 0 3942016 2024-08-23 07:52:58.000000 data.tar.xz\n"}, {"source1": "control.tar.xz", "source2": "control.tar.xz", "unified_diff": null, "details": [{"source1": "control.tar", "source2": "control.tar", "unified_diff": null, "details": [{"source1": "./md5sums", "source2": "./md5sums", "unified_diff": null, "details": [{"source1": "./md5sums", "source2": "./md5sums", "comments": ["Files differ"], "unified_diff": null}]}]}]}, {"source1": "data.tar.xz", "source2": "data.tar.xz", "unified_diff": null, "details": [{"source1": "data.tar", "source2": "data.tar", "unified_diff": null, "details": [{"source1": "./usr/share/doc/python-sqlalchemy-doc/html/changelog/changelog_14.html", "source2": "./usr/share/doc/python-sqlalchemy-doc/html/changelog/changelog_14.html", "unified_diff": "@@ -9239,15 +9239,22 @@\n
See also
\n \nReferences: #4710
\n \n \n-[engine] [change] [performance] [py3k] \u00b6
Disabled the \u201cunicode returns\u201d check that runs on dialect startup when\n+
[engine] [performance] \u00b6
The pool \u201cpre-ping\u201d feature has been refined to not invoke for a DBAPI\n+connection that was just opened in the same checkout operation. pre ping\n+only applies to a DBAPI connection that\u2019s been checked into the pool\n+and is being checked out again.
\n+References: #4524
\n+\n+[engine] [performance] [change] [py3k] \u00b6
Disabled the \u201cunicode returns\u201d check that runs on dialect startup when\n running under Python 3, which for many years has occurred in order to test\n the current DBAPI\u2019s behavior for whether or not it returns Python Unicode\n or Py2K strings for the VARCHAR and NVARCHAR datatypes. The check still\n occurs by default under Python 2, however the mechanism to test the\n behavior will be removed in SQLAlchemy 2.0 when Python 2 support is also\n removed.
\nThis logic was very effective when it was needed, however now that Python 3\n@@ -9258,21 +9265,14 @@\n dialect flags by setting the dialect level flag returns_unicode_strings
\n to one of String.RETURNS_CONDITIONAL
or\n String.RETURNS_BYTES
, both of which will enable Unicode conversion\n even under Python 3.
References: #5315
\n \n[engine] [performance] \u00b6
The pool \u201cpre-ping\u201d feature has been refined to not invoke for a DBAPI\n-connection that was just opened in the same checkout operation. pre ping\n-only applies to a DBAPI connection that\u2019s been checked into the pool\n-and is being checked out again.
\n-References: #4524
\n-\n-[engine] [bug] \u00b6
Revised the Connection.execution_options.schema_translate_map
\n feature such that the processing of the SQL statement to receive a specific\n schema name occurs within the execution phase of the statement, rather than\n at the compile phase. This is to support the statement being efficiently\n cached. Previously, the current schema being rendered into the statement\n for a particular run would be considered as part of the cache key itself,\n meaning that for a run against hundreds of schemas, there would be hundreds\n", "details": [{"source1": "html2text {}", "source2": "html2text {}", "unified_diff": "@@ -6354,15 +6354,21 @@\n returned by the ResultProxy is now the LegacyRow subclass, which maintains\n mapping/tuple hybrid behavior, however the base _\bR_\bo_\bw class now behaves more\n fully like a named tuple.\n See also\n _\bR_\bo_\bw_\bP_\br_\bo_\bx_\by_\b _\bi_\bs_\b _\bn_\bo_\b _\bl_\bo_\bn_\bg_\be_\br_\b _\ba_\b _\b\u201c_\bp_\br_\bo_\bx_\by_\b\u201d_\b;_\b _\bi_\bs_\b _\bn_\bo_\bw_\b _\bc_\ba_\bl_\bl_\be_\bd_\b _\bR_\bo_\bw_\b _\ba_\bn_\bd_\b _\bb_\be_\bh_\ba_\bv_\be_\bs_\b _\bl_\bi_\bk_\be_\b _\ba_\bn_\b _\be_\bn_\bh_\ba_\bn_\bc_\be_\bd\n _\bn_\ba_\bm_\be_\bd_\b _\bt_\bu_\bp_\bl_\be\n References: _\b#_\b4_\b7_\b1_\b0\n-[\b[e\ben\bng\bgi\bin\bne\be]\b] [\b[c\bch\bha\ban\bng\bge\be]\b] [\b[p\bpe\ber\brf\bfo\bor\brm\bma\ban\bnc\bce\be]\b] [\b[p\bpy\by3\b3k\bk]\b] _\b\u00b6\n+[\b[e\ben\bng\bgi\bin\bne\be]\b] [\b[p\bpe\ber\brf\bfo\bor\brm\bma\ban\bnc\bce\be]\b] _\b\u00b6\n+The pool \u201cpre-ping\u201d feature has been refined to not invoke for a DBAPI\n+connection that was just opened in the same checkout operation. pre ping only\n+applies to a DBAPI connection that\u2019s been checked into the pool and is being\n+checked out again.\n+References: _\b#_\b4_\b5_\b2_\b4\n+[\b[e\ben\bng\bgi\bin\bne\be]\b] [\b[p\bpe\ber\brf\bfo\bor\brm\bma\ban\bnc\bce\be]\b] [\b[c\bch\bha\ban\bng\bge\be]\b] [\b[p\bpy\by3\b3k\bk]\b] _\b\u00b6\n Disabled the \u201cunicode returns\u201d check that runs on dialect startup when running\n under Python 3, which for many years has occurred in order to test the current\n DBAPI\u2019s behavior for whether or not it returns Python Unicode or Py2K strings\n for the VARCHAR and NVARCHAR datatypes. The check still occurs by default under\n Python 2, however the mechanism to test the behavior will be removed in\n SQLAlchemy 2.0 when Python 2 support is also removed.\n This logic was very effective when it was needed, however now that Python 3 is\n@@ -6370,20 +6376,14 @@\n datatypes. In the unlikely case that a third party DBAPI does not support this,\n the conversion logic within _\bS_\bt_\br_\bi_\bn_\bg is still available and the third party\n dialect may specify this in its upfront dialect flags by setting the dialect\n level flag returns_unicode_strings to one of String.RETURNS_CONDITIONAL or\n String.RETURNS_BYTES, both of which will enable Unicode conversion even under\n Python 3.\n References: _\b#_\b5_\b3_\b1_\b5\n-[\b[e\ben\bng\bgi\bin\bne\be]\b] [\b[p\bpe\ber\brf\bfo\bor\brm\bma\ban\bnc\bce\be]\b] _\b\u00b6\n-The pool \u201cpre-ping\u201d feature has been refined to not invoke for a DBAPI\n-connection that was just opened in the same checkout operation. pre ping only\n-applies to a DBAPI connection that\u2019s been checked into the pool and is being\n-checked out again.\n-References: _\b#_\b4_\b5_\b2_\b4\n [\b[e\ben\bng\bgi\bin\bne\be]\b] [\b[b\bbu\bug\bg]\b] _\b\u00b6\n Revised the _\bC_\bo_\bn_\bn_\be_\bc_\bt_\bi_\bo_\bn_\b._\be_\bx_\be_\bc_\bu_\bt_\bi_\bo_\bn_\b__\bo_\bp_\bt_\bi_\bo_\bn_\bs_\b._\bs_\bc_\bh_\be_\bm_\ba_\b__\bt_\br_\ba_\bn_\bs_\bl_\ba_\bt_\be_\b__\bm_\ba_\bp feature such that\n the processing of the SQL statement to receive a specific schema name occurs\n within the execution phase of the statement, rather than at the compile phase.\n This is to support the statement being efficiently cached. Previously, the\n current schema being rendered into the statement for a particular run would be\n considered as part of the cache key itself, meaning that for a run against\n"}]}, {"source1": "./usr/share/doc/python-sqlalchemy-doc/html/orm/examples.html", "source2": "./usr/share/doc/python-sqlalchemy-doc/html/orm/examples.html", "comments": ["Ordering differences only"], "unified_diff": "@@ -319,29 +319,29 @@\n \n
Examples illustrating the asyncio engine feature of SQLAlchemy.
\nListing of files:
async_orm_writeonly.py - Illustrates using write only relationships for simpler handling\n+of ORM collections under asyncio.
\n+async_orm.py - Illustrates use of the sqlalchemy.ext.asyncio.AsyncSession
object\n for asynchronous ORM use.
basic.py - Illustrates the asyncio engine / connection interface.
\n+greenlet_orm.py - Illustrates use of the sqlalchemy.ext.asyncio.AsyncSession object\n+for asynchronous ORM use, including the optional run_sync() method.
\nasync_orm_writeonly.py - Illustrates using write only relationships for simpler handling\n-of ORM collections under asyncio.
\n+basic.py - Illustrates the asyncio engine / connection interface.
\ngather_orm_statements.py - Illustrates how to run many statements concurrently using asyncio.gather()
\n along many asyncio database connections, merging ORM results into a single\n AsyncSession
.
greenlet_orm.py - Illustrates use of the sqlalchemy.ext.asyncio.AsyncSession object\n-for asynchronous ORM use, including the optional run_sync() method.
\n-An example of persistence for a directed graph structure. The\n graph is stored as a collection of edges, each referencing both a\n@@ -378,37 +378,37 @@\n subclassing the HasAddresses
mixin, which ensures that the\n parent class is provided with an addresses
collection\n which contains Address
objects.
The discriminator_on_association.py and generic_fk.py scripts\n are modernized versions of recipes presented in the 2007 blog post\n Polymorphic Associations with SQLAlchemy.
\nListing of files:
discriminator_on_association.py - Illustrates a mixin which provides a generic association\n+using a single target table and a single association table,\n+referred to by all parent tables. The association table\n+contains a \u201cdiscriminator\u201d column which determines what type of\n+parent object associates to each particular row in the association\n+table.
\n+table_per_related.py - Illustrates a generic association which persists association\n+objects within individual tables, each one generated to persist\n+those objects on behalf of a particular parent class.
\n+generic_fk.py - Illustrates a so-called \u201cgeneric foreign key\u201d, in a similar fashion\n to that of popular frameworks such as Django, ROR, etc. This\n approach bypasses standard referential integrity\n practices, in that the \u201cforeign key\u201d column is not actually\n constrained to refer to any particular table; instead,\n in-application logic is used to determine which table is referenced.
\ntable_per_association.py - Illustrates a mixin which provides a generic association\n via a individually generated association tables for each parent class.\n The associated objects themselves are persisted in a single table\n shared among all parents.
\ntable_per_related.py - Illustrates a generic association which persists association\n-objects within individual tables, each one generated to persist\n-those objects on behalf of a particular parent class.
\n-discriminator_on_association.py - Illustrates a mixin which provides a generic association\n-using a single target table and a single association table,\n-referred to by all parent tables. The association table\n-contains a \u201cdiscriminator\u201d column which determines what type of\n-parent object associates to each particular row in the association\n-table.
\n-Illustrates the \u201cmaterialized paths\u201d pattern for hierarchical data using the\n SQLAlchemy ORM.
\n@@ -477,33 +477,33 @@\nSee also
\n \nListing of files:
bulk_updates.py - This series of tests will illustrate different ways to UPDATE a large number\n-of rows in bulk (under construction! there\u2019s just one test at the moment)
\n-bulk_inserts.py - This series of tests illustrates different ways to INSERT a large number\n-of rows in bulk.
\n-__main__.py - Allows the examples/performance package to be run as a script.
\n-large_resultsets.py - In this series of tests, we are looking at time to load a large number\n-of very small and simple rows.
\n-single_inserts.py - In this series of tests, we\u2019re looking at a method that inserts a row\n within a distinct transaction, and afterwards returns to essentially a\n \u201cclosed\u201d state. This would be analogous to an API call that starts up\n a database connection, inserts the row, commits and closes.
\nshort_selects.py - This series of tests illustrates different ways to SELECT a single\n record by primary key
\n__main__.py - Allows the examples/performance package to be run as a script.
\n+bulk_inserts.py - This series of tests illustrates different ways to INSERT a large number\n+of rows in bulk.
\n+large_resultsets.py - In this series of tests, we are looking at time to load a large number\n+of very small and simple rows.
\n+bulk_updates.py - This series of tests will illustrate different ways to UPDATE a large number\n+of rows in bulk (under construction! there\u2019s just one test at the moment)
\n+This is the default form of run:
\n$ python -m examples.performance single_inserts\n@@ -751,23 +751,23 @@\n Several examples that illustrate the technique of intercepting changes\n that would be first interpreted as an UPDATE on a row, and instead turning\n it into an INSERT of a new row, leaving the previous row intact as\n a historical version.
\n Compare to the Versioning with a History Table example which writes a\n history row to a separate history table.
\n Listing of files:
versioned_rows_w_versionid.py - Illustrates a method to intercept changes on objects, turning\n-an UPDATE statement on a single row into an INSERT statement, so that a new\n-row is inserted with the new data, keeping the old row intact.
\n-versioned_update_old_row.py - Illustrates the same UPDATE into INSERT technique of versioned_rows.py
,\n but also emits an UPDATE on the old row to affect a change in timestamp.\n Also includes a SessionEvents.do_orm_execute()
hook to limit queries\n to only the most recent version.
versioned_rows_w_versionid.py - Illustrates a method to intercept changes on objects, turning\n+an UPDATE statement on a single row into an INSERT statement, so that a new\n+row is inserted with the new data, keeping the old row intact.
\n+versioned_map.py - A variant of the versioned_rows example built around the\n concept of a \u201cvertical table\u201d structure, like those illustrated in\n Vertical Attribute Mapping examples.
\nversioned_rows.py - Illustrates a method to intercept changes on objects, turning\n an UPDATE statement on a single row into an INSERT statement, so that a new\n row is inserted with the new data, keeping the old row intact.
\n@@ -815,42 +815,42 @@\nWorking examples of single-table, joined-table, and concrete-table\n inheritance as described in Mapping Class Inheritance Hierarchies.
\nListing of files:
concrete.py - Concrete-table (table-per-class) inheritance example.
\n+joined.py - Joined-table (table-per-subclass) inheritance example.
\nsingle.py - Single-table (table-per-hierarchy) inheritance example.
\nconcrete.py - Concrete-table (table-per-class) inheritance example.
\n-Examples illustrating modifications to SQLAlchemy\u2019s attribute management\n system.
\nListing of files:
listen_for_events.py - Illustrates how to attach events to all instrumented attributes\n and listen for change events.
\ncustom_management.py - Illustrates customized class instrumentation, using\n-the sqlalchemy.ext.instrumentation
extension package.
active_column_defaults.py - Illustrates use of the AttributeEvents.init_scalar()
\n event, in conjunction with Core column defaults to provide\n ORM objects that automatically produce the default value\n when an un-set attribute is accessed.
custom_management.py - Illustrates customized class instrumentation, using\n+the sqlalchemy.ext.instrumentation
extension package.
A basic example of using the SQLAlchemy Sharding API.\n Sharding refers to horizontally scaling data across multiple\n@@ -879,24 +879,24 @@\n
The construction of generic sharding routines is an ambitious approach\n to the issue of organizing instances among multiple databases. For a\n more plain-spoken alternative, the \u201cdistinct entity\u201d approach\n is a simple method of assigning objects to different tables (and potentially\n database nodes) in an explicit way - described on the wiki at\n EntityName.
\nListing of files:
separate_schema_translates.py - Illustrates sharding using a single database with multiple schemas,\n+where a different \u201cschema_translates_map\u201d can be used for each shard.
\n+separate_tables.py - Illustrates sharding using a single SQLite database, that will however\n have multiple tables using a naming convention.
\nasyncio.py - Illustrates sharding API used with asyncio.
\nseparate_databases.py - Illustrates sharding using distinct SQLite databases.
\nseparate_schema_translates.py - Illustrates sharding using a single database with multiple schemas,\n-where a different \u201cschema_translates_map\u201d can be used for each shard.
\n-