{"diffoscope-json-version": 1, "source1": "/srv/reproducible-results/rbuild-debian/r-b-build.EE6TIstb/b1/pykafka_2.7.0-2.1_armhf.changes", "source2": "/srv/reproducible-results/rbuild-debian/r-b-build.EE6TIstb/b2/pykafka_2.7.0-2.1_armhf.changes", "unified_diff": null, "details": [{"source1": "Files", "source2": "Files", "unified_diff": "@@ -1,3 +1,3 @@\n \n f55d61a70570ee8c1670d4c58d245ce2 35408 debug optional python3-pykafka-dbgsym_2.7.0-2.1_armhf.deb\n- af74532f1416f7133362006f95834aeb 105032 python optional python3-pykafka_2.7.0-2.1_armhf.deb\n+ 8dc61d5e9dc8da487da1d5910d6ed3f0 105016 python optional python3-pykafka_2.7.0-2.1_armhf.deb\n"}, {"source1": "python3-pykafka_2.7.0-2.1_armhf.deb", "source2": "python3-pykafka_2.7.0-2.1_armhf.deb", "unified_diff": null, "details": [{"source1": "file list", "source2": "file list", "unified_diff": "@@ -1,3 +1,3 @@\n -rw-r--r-- 0 0 0 4 2022-11-12 12:50:12.000000 debian-binary\n--rw-r--r-- 0 0 0 2452 2022-11-12 12:50:12.000000 control.tar.xz\n--rw-r--r-- 0 0 0 102388 2022-11-12 12:50:12.000000 data.tar.xz\n+-rw-r--r-- 0 0 0 2456 2022-11-12 12:50:12.000000 control.tar.xz\n+-rw-r--r-- 0 0 0 102368 2022-11-12 12:50:12.000000 data.tar.xz\n"}, {"source1": "control.tar.xz", "source2": "control.tar.xz", "unified_diff": null, "details": [{"source1": "control.tar", "source2": "control.tar", "unified_diff": null, "details": [{"source1": "./md5sums", "source2": "./md5sums", "unified_diff": null, "details": [{"source1": "./md5sums", "source2": "./md5sums", "comments": ["Files differ"], "unified_diff": null}]}]}]}, {"source1": "data.tar.xz", "source2": "data.tar.xz", "unified_diff": null, "details": [{"source1": "data.tar", "source2": "data.tar", "unified_diff": null, "details": [{"source1": "./usr/lib/python3/dist-packages/.coverage", "source2": "./usr/lib/python3/dist-packages/.coverage", "unified_diff": null, "details": [{"source1": "sqlite3 {} .dump", "source2": "sqlite3 {} .dump", "unified_diff": "@@ -14,15 +14,15 @@\n -- 'has_arcs' boolean -- Is this data recording branches?\n -- 'sys_argv' text -- The coverage command line that recorded the data.\n -- 'version' text -- The version of coverage.py that made the file.\n -- 'when' text -- Datetime when the file was created.\n );\n INSERT INTO meta VALUES('sys_argv','[''/usr/lib/python3/dist-packages/pytest/__main__.py'', ''tests'', ''--ignore=tests/pykafka/rdkafka/test_rd_kafka_consumer.py'', ''--ignore=tests/pykafka/rdkafka/test_simple_consumer.py'', ''--ignore=tests/pykafka/rdkafka/test_ssl.py'', ''--ignore=tests/pykafka/test_balancedconsumer.py'', ''--ignore=tests/pykafka/test_cluster.py'', ''--ignore=tests/pykafka/test_partition.py'', ''--ignore=tests/pykafka/test_producer.py'', ''--ignore=tests/pykafka/test_simpleconsumer.py'', ''--ignore=tests/pykafka/test_ssl.py'', ''--ignore=tests/pykafka/utils/test_compression.py'']');\n INSERT INTO meta VALUES('version','6.5.0');\n-INSERT INTO meta VALUES('when','2022-11-20 17:48:28');\n+INSERT INTO meta VALUES('when','2022-11-21 19:53:37');\n INSERT INTO meta VALUES('has_arcs','0');\n CREATE TABLE file (\n -- A row per file measured.\n id integer primary key,\n path text,\n unique (path)\n );\n@@ -49,20 +49,20 @@\n INSERT INTO file VALUES(21,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/producer.py');\n INSERT INTO file VALUES(22,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/partitioners.py');\n INSERT INTO file VALUES(23,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/rdkafka/helpers.py');\n INSERT INTO file VALUES(24,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/rdkafka/simple_consumer.py');\n INSERT INTO file VALUES(25,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/managedbalancedconsumer.py');\n INSERT INTO file VALUES(26,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/partition.py');\n INSERT INTO file VALUES(27,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/client.py');\n-INSERT INTO file VALUES(28,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/test/utils.py');\n-INSERT INTO file VALUES(29,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/test/kafka_instance.py');\n-INSERT INTO file VALUES(30,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/test/__init__.py');\n-INSERT INTO file VALUES(31,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/cli/kafka_tools.py');\n-INSERT INTO file VALUES(32,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/cli/__init__.py');\n-INSERT INTO file VALUES(33,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/cli/__main__.py');\n+INSERT INTO file VALUES(28,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/cli/__init__.py');\n+INSERT INTO file VALUES(29,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/cli/kafka_tools.py');\n+INSERT INTO file VALUES(30,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/cli/__main__.py');\n+INSERT INTO file VALUES(31,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/test/utils.py');\n+INSERT INTO file VALUES(32,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/test/kafka_instance.py');\n+INSERT INTO file VALUES(33,'/build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka/test/__init__.py');\n CREATE TABLE context (\n -- A row per context measured.\n id integer primary key,\n context text,\n unique (context)\n );\n INSERT INTO context VALUES(1,'');\n"}]}, {"source1": "./usr/lib/python3/dist-packages/coverage.xml", "source2": "./usr/lib/python3/dist-packages/coverage.xml", "unified_diff": null, "details": [{"source1": "./usr/lib/python3/dist-packages/coverage.xml", "source2": "./usr/lib/python3/dist-packages/coverage.xml", "unified_diff": "@@ -1,9 +1,9 @@\n \n-\n+\n \n \n \n /build/pykafka-2.7.0/.pybuild/cpython3_3.10_pykafka/build/pykafka\n \n \n \n"}]}]}]}]}]}