{"diffoscope-json-version": 1, "source1": "/srv/reproducible-results/rbuild-debian/r-b-build.RQ6UGCG8/b1/python-xarray_2025.01.2-1_i386.changes", "source2": "/srv/reproducible-results/rbuild-debian/r-b-build.RQ6UGCG8/b2/python-xarray_2025.01.2-1_i386.changes", "unified_diff": null, "details": [{"source1": "Files", "source2": "Files", "unified_diff": "@@ -1,3 +1,3 @@\n \n- 7aa1a6684415050ee6e5a8039db05ff1 2007888 doc optional python-xarray-doc_2025.01.2-1_all.deb\n+ 5cc999a1bc7c7e66c91677fff73b5245 2007716 doc optional python-xarray-doc_2025.01.2-1_all.deb\n 65b17ea58f6b0a83ea67cd4000a96d9f 799856 python optional python3-xarray_2025.01.2-1_all.deb\n"}, {"source1": "python-xarray-doc_2025.01.2-1_all.deb", "source2": "python-xarray-doc_2025.01.2-1_all.deb", "unified_diff": null, "details": [{"source1": "file list", "source2": "file list", "unified_diff": "@@ -1,3 +1,3 @@\n -rw-r--r-- 0 0 0 4 2025-02-02 11:36:57.000000 debian-binary\n--rw-r--r-- 0 0 0 6172 2025-02-02 11:36:57.000000 control.tar.xz\n--rw-r--r-- 0 0 0 2001524 2025-02-02 11:36:57.000000 data.tar.xz\n+-rw-r--r-- 0 0 0 6168 2025-02-02 11:36:57.000000 control.tar.xz\n+-rw-r--r-- 0 0 0 2001356 2025-02-02 11:36:57.000000 data.tar.xz\n"}, {"source1": "control.tar.xz", "source2": "control.tar.xz", "unified_diff": null, "details": [{"source1": "control.tar", "source2": "control.tar", "unified_diff": null, "details": [{"source1": "./control", "source2": "./control", "unified_diff": "@@ -1,13 +1,13 @@\n Package: python-xarray-doc\n Source: python-xarray\n Version: 2025.01.2-1\n Architecture: all\n Maintainer: Debian Science Maintainers Visualizing your datasets is quick and convenient: Note the automatic labeling with names and units. Our effort in adding metadata attributes has paid off! Many aspects of these figures are customizable: see Plotting.Plotting\u00b6
\n In [37]: data.plot()\n-Out[37]: <matplotlib.collections.QuadMesh at 0xe8531818>\n+Out[37]: <matplotlib.collections.QuadMesh at 0xea02d818>\n
\n
pandas\u00b6
\n", "details": [{"source1": "html2text {}", "source2": "html2text {}", "unified_diff": "@@ -253,15 +253,15 @@\n [0.37342613, 1.49497537, 1.33584385]])\n Coordinates:\n * x (x) int32 8B 10 20\n Dimensions without coordinates: y\n *\b**\b**\b**\b**\b* P\bPl\blo\bot\btt\bti\bin\bng\bg_\b?\b\u00b6 *\b**\b**\b**\b**\b*\n Visualizing your datasets is quick and convenient:\n In [37]: data.plot()\n-Out[37]: apply_ufunc
\", \"Compare weighted and unweighted mean temperature\", \"Blank template\", \"Calculating Seasonal Averages from Time Series of Monthly Means\", \"Working with Multidimensional Coordinates\", \"Visualization Gallery\", \"Toy weather data\", \"Gallery\", \"Frequently Asked Questions\", \"Getting Started\", \"Installation\", \"Quick overview\", \"Overview: Why xarray?\", \"Getting Help\", \"How do I \\u2026\", \"Xarray documentation\", \"Alternative chunked array types\", \"Integrating with duck arrays\", \"Extending xarray using accessors\", \"How to add a new backend\", \"How to create a custom index\", \"Xarray Internals\", \"Internal Design\", \"Interoperability of Xarray\", \"Zarr Encoding Specification\", \"Development roadmap\", \"Tutorials and Videos\", \"Combining data\", \"Computation\", \"Parallel Computing with Dask\", \"Data Structures\", \"Working with numpy-like arrays\", \"GroupBy: Group and Bin Data\", \"Hierarchical data\", \"User Guide\", \"Indexing and selecting data\", \"Interpolating data\", \"Reading and writing files\", \"Configuration\", \"Working with pandas\", \"Plotting\", \"Reshaping and reorganizing data\", \"Terminology\", \"Testing your code\", \"Time series data\", \"Weather and climate data\", \"What\\u2019s New\"],\n \"titleterms\": {\n \"\": [13, 54],\n \"0\": 54,\n \"01\": 54,\n \"02\": 54,\n"}]}, {"source1": "./usr/share/doc/python-xarray-doc/html/user-guide/data-structures.html", "source2": "./usr/share/doc/python-xarray-doc/html/user-guide/data-structures.html", "unified_diff": "@@ -731,18 +731,18 @@\n a method call with an external function (e.g., ds.pipe(func)
) instead of\n simply calling it (e.g., func(ds)
). This allows you to write pipelines for\n transforming your data (using \u201cmethod chaining\u201d) instead of writing hard to\n follow nested function calls:
# these lines are equivalent, but with pipe we can make the logic flow\n # entirely from left to right\n In [64]: plt.plot((2 * ds.temperature.sel(loc=0)).mean("instrument"))\n-Out[64]: [<matplotlib.lines.Line2D at 0xe8749500>]\n+Out[64]: [<matplotlib.lines.Line2D at 0xea24a500>]\n \n In [65]: (ds.temperature.sel(loc=0).pipe(lambda x: 2 * x).mean("instrument").pipe(plt.plot))\n-Out[65]: [<matplotlib.lines.Line2D at 0xe455ef50>]\n+Out[65]: [<matplotlib.lines.Line2D at 0xe4068f50>]\n
Both pipe
and assign
replicate the pandas methods of the same names\n (DataFrame.pipe
and\n DataFrame.assign
).
With xarray, there is no performance penalty for creating new datasets, even if\n variables are lazily loaded from a file on disk. Creating new objects instead\n", "details": [{"source1": "html2text {}", "source2": "html2text {}", "unified_diff": "@@ -585,19 +585,19 @@\n There is also the pipe() method that allows you to use a method call with an\n external function (e.g., ds.pipe(func)) instead of simply calling it (e.g.,\n func(ds)). This allows you to write pipelines for transforming your data (using\n \u201cmethod chaining\u201d) instead of writing hard to follow nested function calls:\n # these lines are equivalent, but with pipe we can make the logic flow\n # entirely from left to right\n In [64]: plt.plot((2 * ds.temperature.sel(loc=0)).mean(\"instrument\"))\n-Out[64]: [ If you were a previous user of the prototype xarray-contrib/datatree package, this is different from what you\u2019re used to!\n In that package the data model was that the data stored in each node actually was completely unrelated. The data model is now slightly stricter.\n This allows us to provide features like Coordinate Inheritance. To demonstrate, let\u2019s first generate some example datasets which are not aligned with one another: Now we have a valid This is a useful way to organise our data because we can still operate on all the groups at once.\n For example we can extract all three timeseries at a specific lat-lon location: or compute the standard deviation of each timeseries to find out how it varies with sampling frequency: This helps to differentiate which variables are defined on the datatree node that you are currently looking at, and which were defined somewhere above it. We can also still perform all the same operations on the whole tree:# (drop the attributes just to make the printed representation shorter)\n In [89]: ds = xr.tutorial.open_dataset("air_temperature").drop_attrs()\n-PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/first-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n+PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/second-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n \n \n In [90]: ds_daily = ds.resample(time="D").mean("time")\n KeyError: "No variable named 'time'. Variables on the dataset include ['foo', 'x', 'letters']"\n \n \n In [91]: ds_weekly = ds.resample(time="W").mean("time")\n@@ -1054,15 +1054,15 @@\n \u2514\u2500\u2500 Group: /b/B\n
DataTree
structure which contains all the data at each different time frequency, stored in a separate group.In [100]: dt.sel(lat=75, lon=300)\n-ValueError: Dimensions {'lat', 'lon'} do not exist. Expected one or more of set()\n+ValueError: Dimensions {'lon', 'lat'} do not exist. Expected one or more of set()\n
In [101]: dt.std(dim="time")\n ValueError: Dimension(s) 'time' do not exist. Expected one or more of set()\n
In [107]: print(dt["/daily"])\n KeyError: 'Could not find node at /daily'\n
In [108]: dt.sel(lat=[75], lon=[300])\n-ValueError: Dimensions {'lat', 'lon'} do not exist. Expected one or more of set()\n+ValueError: Dimensions {'lon', 'lat'} do not exist. Expected one or more of set()\n \n \n In [109]: dt.std(dim="time")\n ValueError: Dimension(s) 'time' do not exist. Expected one or more of set()\n
In [52]: ds = xr.tutorial.open_dataset("air_temperature")\n-PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/first-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n+PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/second-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n \n \n # Define target latitude and longitude (where weather stations might be)\n In [53]: target_lon = xr.DataArray([200, 201, 202, 205], dims="points")\n \n In [54]: target_lat = xr.DataArray([31, 41, 42, 42], dims="points")\n \n@@ -697,15 +697,15 @@\n
To select and assign values to a portion of a DataArray()
you\n can use indexing with .loc
:
In [57]: ds = xr.tutorial.open_dataset("air_temperature")\n-PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/first-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n+PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/second-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n \n \n # add an empty 2D dataarray\n In [58]: ds["empty"] = xr.full_like(ds.air.mean("time"), fill_value=0)\n AttributeError: 'Dataset' object has no attribute 'air'\n \n \n@@ -869,15 +869,15 @@\n
You can also assign values to all variables of a Dataset
at once:
In [83]: ds_org = xr.tutorial.open_dataset("eraint_uvz").isel(\n ....: latitude=slice(56, 59), longitude=slice(255, 258), level=0\n ....: )\n ....: \n-PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/first-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n+PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/second-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n \n \n # set all values to 0\n In [84]: ds = xr.zeros_like(ds_org)\n NameError: name 'ds_org' is not defined\n \n \n", "details": [{"source1": "html2text {}", "source2": "html2text {}", "unified_diff": "@@ -474,15 +474,15 @@\n collection specified weather station latitudes and longitudes. To trigger\n vectorized indexing behavior you will need to provide the selection dimensions\n with a new shared output dimension name. In the example below, the selections\n of the closest latitude and longitude are renamed to an output dimension named\n \u201cpoints\u201d:\n In [52]: ds = xr.tutorial.open_dataset(\"air_temperature\")\n PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not\n-create data cache folder '/nonexistent/first-build/.cache/\n+create data cache folder '/nonexistent/second-build/.cache/\n xarray_tutorial_data'. Will not be able to download data files.\n \n \n # Define target latitude and longitude (where weather stations might be)\n In [53]: target_lon = xr.DataArray([200, 201, 202, 205], dims=\"points\")\n \n In [54]: target_lat = xr.DataArray([31, 41, 42, 42], dims=\"points\")\n@@ -513,15 +513,15 @@\n selected subpart of the target array (except for the explicitly indexed\n dimensions with .loc/.sel). Otherwise, IndexError will be raised.\n *\b**\b**\b**\b**\b* A\bAs\bss\bsi\big\bgn\bni\bin\bng\bg v\bva\bal\blu\bue\bes\bs w\bwi\bit\bth\bh i\bin\bnd\bde\bex\bxi\bin\bng\bg_\b?\b\u00b6 *\b**\b**\b**\b**\b*\n To select and assign values to a portion of a DataArray() you can use indexing\n with .loc :\n In [57]: ds = xr.tutorial.open_dataset(\"air_temperature\")\n PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not\n-create data cache folder '/nonexistent/first-build/.cache/\n+create data cache folder '/nonexistent/second-build/.cache/\n xarray_tutorial_data'. Will not be able to download data files.\n \n \n # add an empty 2D dataarray\n In [58]: ds[\"empty\"] = xr.full_like(ds.air.mean(\"time\"), fill_value=0)\n AttributeError: 'Dataset' object has no attribute 'air'\n \n@@ -673,15 +673,15 @@\n Dimensions without coordinates: x\n You can also assign values to all variables of a Dataset at once:\n In [83]: ds_org = xr.tutorial.open_dataset(\"eraint_uvz\").isel(\n ....: latitude=slice(56, 59), longitude=slice(255, 258), level=0\n ....: )\n ....:\n PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not\n-create data cache folder '/nonexistent/first-build/.cache/\n+create data cache folder '/nonexistent/second-build/.cache/\n xarray_tutorial_data'. Will not be able to download data files.\n \n \n # set all values to 0\n In [84]: ds = xr.zeros_like(ds_org)\n NameError: name 'ds_org' is not defined\n \n"}]}, {"source1": "./usr/share/doc/python-xarray-doc/html/user-guide/interpolation.html", "source2": "./usr/share/doc/python-xarray-doc/html/user-guide/interpolation.html", "unified_diff": "@@ -237,24 +237,24 @@\n ....: np.sin(np.linspace(0, 2 * np.pi, 10)),\n ....: dims="x",\n ....: coords={"x": np.linspace(0, 1, 10)},\n ....: )\n ....: \n \n In [17]: da.plot.line("o", label="original")\n-Out[17]: [<matplotlib.lines.Line2D at 0xe3001450>]\n+Out[17]: [<matplotlib.lines.Line2D at 0xe2adf450>]\n \n In [18]: da.interp(x=np.linspace(0, 1, 100)).plot.line(label="linear (default)")\n-Out[18]: [<matplotlib.lines.Line2D at 0xe308bea0>]\n+Out[18]: [<matplotlib.lines.Line2D at 0xe2b69ea0>]\n \n In [19]: da.interp(x=np.linspace(0, 1, 100), method="cubic").plot.line(label="cubic")\n-Out[19]: [<matplotlib.lines.Line2D at 0xe8535190>]\n+Out[19]: [<matplotlib.lines.Line2D at 0xea031190>]\n \n In [20]: plt.legend()\n-Out[20]: <matplotlib.legend.Legend at 0xe8531e90>\n+Out[20]: <matplotlib.legend.Legend at 0xea02de90>\n
Additional keyword arguments can be passed to scipy\u2019s functions.
\n# fill 0 for the outside of the original coordinates.\n In [21]: da.interp(x=np.linspace(-0.5, 1.5, 10), kwargs={"fill_value": 0.0})\n@@ -439,15 +439,15 @@\n see Missing values.\n \n \n Example\u00b6
\n Let\u2019s see how interp()
works on real data.
\n # Raw data\n In [44]: ds = xr.tutorial.open_dataset("air_temperature").isel(time=0)\n-PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/first-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n+PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/second-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n \n \n In [45]: fig, axes = plt.subplots(ncols=2, figsize=(10, 4))\n \n In [46]: ds.air.plot(ax=axes[0])\n AttributeError: 'Dataset' object has no attribute 'air'\n \n@@ -511,15 +511,15 @@\n ....: axes[0].plot(*xr.broadcast(lon.isel(z=idx), lat.isel(z=idx)), "--k")\n ....: \n \n In [61]: axes[0].set_title("Raw data")\n Out[61]: Text(0.5, 1.0, 'Raw data')\n \n In [62]: dsi = ds.interp(lon=lon, lat=lat)\n-ValueError: Dimensions {'lat', 'lon'} do not exist. Expected one or more of FrozenMappingWarningOnValuesAccess({'x': 3, 'y': 4})\n+ValueError: Dimensions {'lon', 'lat'} do not exist. Expected one or more of FrozenMappingWarningOnValuesAccess({'x': 3, 'y': 4})\n \n \n In [63]: dsi.air.plot(ax=axes[1])\n NameError: name 'dsi' is not defined\n \n \n In [64]: axes[1].set_title("Remapped data")\n", "details": [{"source1": "html2text {}", "source2": "html2text {}", "unified_diff": "@@ -154,26 +154,26 @@\n ....: np.sin(np.linspace(0, 2 * np.pi, 10)),\n ....: dims=\"x\",\n ....: coords={\"x\": np.linspace(0, 1, 10)},\n ....: )\n ....:\n \n In [17]: da.plot.line(\"o\", label=\"original\")\n-Out[17]: []\n+Out[17]: []\n \n In [18]: da.interp(x=np.linspace(0, 1, 100)).plot.line(label=\"linear\n (default)\")\n-Out[18]: []\n+Out[18]: []\n \n In [19]: da.interp(x=np.linspace(0, 1, 100), method=\"cubic\").plot.line\n (label=\"cubic\")\n-Out[19]: []\n+Out[19]: []\n \n In [20]: plt.legend()\n-Out[20]: \n+Out[20]: \n _\b[_\b__\bb_\bu_\bi_\bl_\bd_\b/_\bh_\bt_\bm_\bl_\b/_\b__\bs_\bt_\ba_\bt_\bi_\bc_\b/_\bi_\bn_\bt_\be_\br_\bp_\bo_\bl_\ba_\bt_\bi_\bo_\bn_\b__\bs_\ba_\bm_\bp_\bl_\be_\b1_\b._\bp_\bn_\bg_\b]\n Additional keyword arguments can be passed to scipy\u2019s functions.\n # fill 0 for the outside of the original coordinates.\n In [21]: da.interp(x=np.linspace(-0.5, 1.5, 10), kwargs={\"fill_value\": 0.0})\n Out[21]:\n Size: 80B\n array([ 0. , 0. , 0. , 0.814, 0.604, -0.604, -0.814, 0. , 0. ,\n@@ -337,15 +337,15 @@\n * x (x) float64 24B 0.5 1.5 2.5\n For the details of interpolate_na(), see _\bM_\bi_\bs_\bs_\bi_\bn_\bg_\b _\bv_\ba_\bl_\bu_\be_\bs.\n *\b**\b**\b**\b**\b* E\bEx\bxa\bam\bmp\bpl\ble\be_\b?\b\u00b6 *\b**\b**\b**\b**\b*\n Let\u2019s see how interp() works on real data.\n # Raw data\n In [44]: ds = xr.tutorial.open_dataset(\"air_temperature\").isel(time=0)\n PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not\n-create data cache folder '/nonexistent/first-build/.cache/\n+create data cache folder '/nonexistent/second-build/.cache/\n xarray_tutorial_data'. Will not be able to download data files.\n \n \n In [45]: fig, axes = plt.subplots(ncols=2, figsize=(10, 4))\n \n In [46]: ds.air.plot(ax=axes[0])\n AttributeError: 'Dataset' object has no attribute 'air'\n@@ -410,15 +410,15 @@\n k\")\n ....:\n \n In [61]: axes[0].set_title(\"Raw data\")\n Out[61]: Text(0.5, 1.0, 'Raw data')\n \n In [62]: dsi = ds.interp(lon=lon, lat=lat)\n-ValueError: Dimensions {'lat', 'lon'} do not exist. Expected one or more of\n+ValueError: Dimensions {'lon', 'lat'} do not exist. Expected one or more of\n FrozenMappingWarningOnValuesAccess({'x': 3, 'y': 4})\n \n \n In [63]: dsi.air.plot(ax=axes[1])\n NameError: name 'dsi' is not defined\n \n \n"}]}, {"source1": "./usr/share/doc/python-xarray-doc/html/user-guide/io.html", "source2": "./usr/share/doc/python-xarray-doc/html/user-guide/io.html", "unified_diff": "@@ -630,15 +630,15 @@\n ....: "y": pd.date_range("2000-01-01", periods=5),\n ....: "z": ("x", list("abcd")),\n ....: },\n ....: )\n ....: \n \n In [13]: ds.to_zarr("path/to/directory.zarr")\n-Out[13]: <xarray.backends.zarr.ZarrStore at 0xe2e72cd0>\n+Out[13]: <xarray.backends.zarr.ZarrStore at 0xe291e388>\n
\n \n (The suffix .zarr
is optional\u2013just a reminder that a zarr store lives\n there.) If the directory does not exist, it will be created. If a zarr\n store is already present at that path, an error will be raised, preventing it\n from being overwritten. To override this behavior and overwrite an existing\n store, add mode='w'
when invoking to_zarr()
.
\n@@ -724,36 +724,36 @@\n \n In [18]: ds = xr.Dataset({"foo": ("x", dummies)}, coords={"x": np.arange(30)})\n \n In [19]: path = "path/to/directory.zarr"\n \n # Now we write the metadata without computing any array values\n In [20]: ds.to_zarr(path, compute=False)\n-Out[20]: Delayed('_finalize_store-38ad501d-0483-463e-a35a-dc8139172eda')\n+Out[20]: Delayed('_finalize_store-c135d9f5-8967-49cb-beb2-f02e92ce7636')\n
Now, a Zarr store with the correct variable shapes and attributes exists that\n can be filled out by subsequent calls to to_zarr
.\n Setting region="auto"
will open the existing store and determine the\n correct alignment of the new data with the existing dimensions, or as an\n explicit mapping from dimension names to Python slice
objects indicating\n where the data should be written (in index space, not label space), e.g.,
# For convenience, we'll slice a single dataset, but in the real use-case\n # we would create them separately possibly even from separate processes.\n In [21]: ds = xr.Dataset({"foo": ("x", np.arange(30))}, coords={"x": np.arange(30)})\n \n # Any of the following region specifications are valid\n In [22]: ds.isel(x=slice(0, 10)).to_zarr(path, region="auto")\n-Out[22]: <xarray.backends.zarr.ZarrStore at 0xe29dde80>\n+Out[22]: <xarray.backends.zarr.ZarrStore at 0xe24b9df0>\n \n In [23]: ds.isel(x=slice(10, 20)).to_zarr(path, region={"x": "auto"})\n-Out[23]: <xarray.backends.zarr.ZarrStore at 0xe29dd388>\n+Out[23]: <xarray.backends.zarr.ZarrStore at 0xe24b9220>\n \n In [24]: ds.isel(x=slice(20, 30)).to_zarr(path, region={"x": slice(20, 30)})\n-Out[24]: <xarray.backends.zarr.ZarrStore at 0xe29d77c0>\n+Out[24]: <xarray.backends.zarr.ZarrStore at 0xe24b9898>\n
Concurrent writes with region
are safe as long as they modify distinct\n chunks in the underlying Zarr arrays (or use an appropriate lock
).
As a safety check to make it harder to inadvertently override existing values,\n if you set region
then all variables included in a Dataset must have\n dimensions included in region
. Other variables (typically coordinates)\n@@ -769,15 +769,15 @@\n
In [25]: import zarr\n \n In [26]: from numcodecs.blosc import Blosc\n \n In [27]: compressor = Blosc(cname="zstd", clevel=3, shuffle=2)\n \n In [28]: ds.to_zarr("foo.zarr", encoding={"foo": {"compressor": compressor}})\n-Out[28]: <xarray.backends.zarr.ZarrStore at 0xe29e4970>\n+Out[28]: <xarray.backends.zarr.ZarrStore at 0xe24c09b8>\n
Note
\nNot all native zarr compression and filtering options have been tested with\n xarray.
\nChunk sizes may be specified in one of three ways when writing to a zarr store:
\nFor example, let\u2019s say we\u2019re working with a dataset with dimensions\n ('time', 'x', 'y')
, a variable Tair
which is chunked in x
and y
,\n and two multi-dimensional coordinates xc
and yc
:
In [33]: ds = xr.tutorial.open_dataset("rasm")\n-PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/first-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n+PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/second-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n \n \n In [34]: ds["Tair"] = ds["Tair"].chunk({"x": 100, "y": 100})\n KeyError: "No variable named 'Tair'. Variables on the dataset include ['foo', 'x']"\n \n \n In [35]: ds\n@@ -882,15 +882,15 @@\n foo (x) int32 120B 0 1 2 3 4 5 6 7 8 9 ... 21 22 23 24 25 26 27 28 29\n
These multi-dimensional coordinates are only two-dimensional and take up very little\n space on disk or in memory, yet when writing to disk the default zarr behavior is to\n split them into chunks:
\nIn [36]: ds.to_zarr("path/to/directory.zarr", mode="w")\n-Out[36]: <xarray.backends.zarr.ZarrStore at 0xe29f7a90>\n+Out[36]: <xarray.backends.zarr.ZarrStore at 0xe24d3ad8>\n \n In [37]: ! ls -R path/to/directory.zarr\n path/to/directory.zarr:\n foo x\n \n path/to/directory.zarr/foo:\n 0\n@@ -1062,15 +1062,15 @@\n Ncdata\u00b6
\n Ncdata provides more sophisticated means of transferring data, including entire\n datasets. It uses the file saving and loading functions in both projects to provide a\n more \u201ccorrect\u201d translation between them, but still with very low overhead and not\n using actual disk files.
\n For example:
\n In [48]: ds = xr.tutorial.open_dataset("air_temperature_gradient")\n-PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/first-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n+PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/second-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n \n \n In [49]: cubes = ncdata.iris_xarray.cubes_from_xarray(ds)\n NameError: name 'ncdata' is not defined\n \n \n In [50]: print(cubes)\n", "details": [{"source1": "html2text {}", "source2": "html2text {}", "unified_diff": "@@ -481,15 +481,15 @@\n ....: \"y\": pd.date_range(\"2000-01-01\", periods=5),\n ....: \"z\": (\"x\", list(\"abcd\")),\n ....: },\n ....: )\n ....:\n \n In [13]: ds.to_zarr(\"path/to/directory.zarr\")\n-Out[13]: \n+Out[13]: \n (The suffix .zarr is optional\u2013just a reminder that a zarr store lives there.)\n If the directory does not exist, it will be created. If a zarr store is already\n present at that path, an error will be raised, preventing it from being\n overwritten. To override this behavior and overwrite an existing store, add\n mode='w' when invoking to_zarr().\n DataArrays can also be saved to disk using the DataArray.to_zarr() method, and\n loaded from disk using the open_dataarray() function with engine='zarr'.\n@@ -562,35 +562,35 @@\n \n In [18]: ds = xr.Dataset({\"foo\": (\"x\", dummies)}, coords={\"x\": np.arange(30)})\n \n In [19]: path = \"path/to/directory.zarr\"\n \n # Now we write the metadata without computing any array values\n In [20]: ds.to_zarr(path, compute=False)\n-Out[20]: Delayed('_finalize_store-38ad501d-0483-463e-a35a-dc8139172eda')\n+Out[20]: Delayed('_finalize_store-c135d9f5-8967-49cb-beb2-f02e92ce7636')\n Now, a Zarr store with the correct variable shapes and attributes exists that\n can be filled out by subsequent calls to to_zarr. Setting region=\"auto\" will\n open the existing store and determine the correct alignment of the new data\n with the existing dimensions, or as an explicit mapping from dimension names to\n Python slice objects indicating where the data should be written (in index\n space, not label space), e.g.,\n # For convenience, we'll slice a single dataset, but in the real use-case\n # we would create them separately possibly even from separate processes.\n In [21]: ds = xr.Dataset({\"foo\": (\"x\", np.arange(30))}, coords={\"x\": np.arange\n (30)})\n \n # Any of the following region specifications are valid\n In [22]: ds.isel(x=slice(0, 10)).to_zarr(path, region=\"auto\")\n-Out[22]: \n+Out[22]: \n \n In [23]: ds.isel(x=slice(10, 20)).to_zarr(path, region={\"x\": \"auto\"})\n-Out[23]: \n+Out[23]: \n \n In [24]: ds.isel(x=slice(20, 30)).to_zarr(path, region={\"x\": slice(20, 30)})\n-Out[24]: \n+Out[24]: \n Concurrent writes with region are safe as long as they modify distinct chunks\n in the underlying Zarr arrays (or use an appropriate lock).\n As a safety check to make it harder to inadvertently override existing values,\n if you set region then a\bal\bll\bl variables included in a Dataset must have dimensions\n included in region. Other variables (typically coordinates) need to be\n explicitly dropped and/or written in a separate calls to to_zarr with mode='a'.\n *\b**\b**\b**\b* Z\bZa\bar\brr\br C\bCo\bom\bmp\bpr\bre\bes\bss\bso\bor\brs\bs a\ban\bnd\bd F\bFi\bil\blt\bte\ber\brs\bs_\b?\b\u00b6 *\b**\b**\b**\b*\n@@ -601,15 +601,15 @@\n In [25]: import zarr\n \n In [26]: from numcodecs.blosc import Blosc\n \n In [27]: compressor = Blosc(cname=\"zstd\", clevel=3, shuffle=2)\n \n In [28]: ds.to_zarr(\"foo.zarr\", encoding={\"foo\": {\"compressor\": compressor}})\n-Out[28]: \n+Out[28]: \n Note\n Not all native zarr compression and filtering options have been tested with\n xarray.\n *\b**\b**\b**\b* M\bMo\bod\bdi\bif\bfy\byi\bin\bng\bg e\bex\bxi\bis\bst\bti\bin\bng\bg Z\bZa\bar\brr\br s\bst\bto\bor\bre\bes\bs_\b?\b\u00b6 *\b**\b**\b**\b*\n Xarray supports several ways of incrementally writing variables to a Zarr\n store. These options are useful for scenarios when it is infeasible or\n undesirable to write your entire dataset at once.\n@@ -635,28 +635,28 @@\n ....: \"y\": [1, 2, 3, 4, 5],\n ....: \"t\": pd.date_range(\"2001-01-01\", periods=2),\n ....: },\n ....: )\n ....:\n \n In [30]: ds1.to_zarr(\"path/to/directory.zarr\")\n-Out[30]: \n+Out[30]: \n \n In [31]: ds2 = xr.Dataset(\n ....: {\"foo\": ((\"x\", \"y\", \"t\"), np.random.rand(4, 5, 2))},\n ....: coords={\n ....: \"x\": [10, 20, 30, 40],\n ....: \"y\": [1, 2, 3, 4, 5],\n ....: \"t\": pd.date_range(\"2001-01-03\", periods=2),\n ....: },\n ....: )\n ....:\n \n In [32]: ds2.to_zarr(\"path/to/directory.zarr\", append_dim=\"t\")\n-Out[32]: \n+Out[32]: \n *\b**\b**\b**\b* S\bSp\bpe\bec\bci\bif\bfy\byi\bin\bng\bg c\bch\bhu\bun\bnk\bks\bs i\bin\bn a\ba z\bza\bar\brr\br s\bst\bto\bor\bre\be_\b?\b\u00b6 *\b**\b**\b**\b*\n Chunk sizes may be specified in one of three ways when writing to a zarr store:\n 1. Manual chunk sizing through the use of the encoding argument in\n Dataset.to_zarr():\n 2. Automatic chunking based on chunks in dask arrays\n 3. Default chunk behavior determined by the zarr library\n The resulting chunks will be determined based on the order of the above list;\n@@ -675,15 +675,15 @@\n positional ordering of the dimensions in each array. Watch out for arrays with\n differently-ordered dimensions within a single Dataset.\n For example, let\u2019s say we\u2019re working with a dataset with dimensions ('time',\n 'x', 'y'), a variable Tair which is chunked in x and y, and two multi-\n dimensional coordinates xc and yc:\n In [33]: ds = xr.tutorial.open_dataset(\"rasm\")\n PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not\n-create data cache folder '/nonexistent/first-build/.cache/\n+create data cache folder '/nonexistent/second-build/.cache/\n xarray_tutorial_data'. Will not be able to download data files.\n \n \n In [34]: ds[\"Tair\"] = ds[\"Tair\"].chunk({\"x\": 100, \"y\": 100})\n KeyError: \"No variable named 'Tair'. Variables on the dataset include ['foo',\n 'x']\"\n \n@@ -696,15 +696,15 @@\n * x (x) int32 120B 0 1 2 3 4 5 6 7 8 9 ... 21 22 23 24 25 26 27 28 29\n Data variables:\n foo (x) int32 120B 0 1 2 3 4 5 6 7 8 9 ... 21 22 23 24 25 26 27 28 29\n These multi-dimensional coordinates are only two-dimensional and take up very\n little space on disk or in memory, yet when writing to disk the default zarr\n behavior is to split them into chunks:\n In [36]: ds.to_zarr(\"path/to/directory.zarr\", mode=\"w\")\n-Out[36]: \n+Out[36]: \n \n In [37]: ! ls -R path/to/directory.zarr\n path/to/directory.zarr:\n foo x\n \n path/to/directory.zarr/foo:\n 0\n@@ -850,15 +850,15 @@\n _\bN_\bc_\bd_\ba_\bt_\ba provides more sophisticated means of transferring data, including entire\n datasets. It uses the file saving and loading functions in both projects to\n provide a more \u201ccorrect\u201d translation between them, but still with very low\n overhead and not using actual disk files.\n For example:\n In [48]: ds = xr.tutorial.open_dataset(\"air_temperature_gradient\")\n PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not\n-create data cache folder '/nonexistent/first-build/.cache/\n+create data cache folder '/nonexistent/second-build/.cache/\n xarray_tutorial_data'. Will not be able to download data files.\n \n \n In [49]: cubes = ncdata.iris_xarray.cubes_from_xarray(ds)\n NameError: name 'ncdata' is not defined\n \n \n"}]}, {"source1": "./usr/share/doc/python-xarray-doc/html/user-guide/plotting.html", "source2": "./usr/share/doc/python-xarray-doc/html/user-guide/plotting.html", "unified_diff": "@@ -100,15 +100,15 @@\n In [3]: import matplotlib.pyplot as plt\n \n In [4]: import xarray as xr\n
\n \n For these examples we\u2019ll use the North American air temperature dataset.
\n In [5]: airtemps = xr.tutorial.open_dataset("air_temperature")\n-PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/first-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n+PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/second-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n \n \n In [6]: airtemps\n NameError: name 'airtemps' is not defined\n \n \n # Convert to celsius\n@@ -445,15 +445,15 @@\n \n # Apply a nonlinear transformation to one of the coords\n In [50]: b.coords["lat"] = np.log(b.coords["lat"])\n KeyError: 'lat'\n \n \n In [51]: b.plot()\n-Out[51]: [<matplotlib.lines.Line2D at 0xe2153500>]\n+Out[51]: [<matplotlib.lines.Line2D at 0xe0bcd500>]\n
\n \n
\n \n \n \n Other types of plot\u00b6
\n@@ -857,117 +857,117 @@\n * y (y) float64 88B 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0\n * z (z) int32 16B 0 1 2 3\n * w (w) <U5 80B 'one' 'two' 'three' 'five'\n Attributes:\n units: Aunits\n \n In [99]: ds.A.plot.scatter(x="y")\n-Out[99]: <matplotlib.collections.PathCollection at 0xe2eeb3c8>\n+Out[99]: <matplotlib.collections.PathCollection at 0xe29c73c8>\n
Same plot can be displayed using the dataset:
\nIn [100]: ds.plot.scatter(x="y", y="A")\n-Out[100]: <matplotlib.collections.PathCollection at 0xe0da5df0>\n+Out[100]: <matplotlib.collections.PathCollection at 0xe0840df0>\n
Now suppose we want to scatter the A
DataArray against the B
DataArray
In [101]: ds.plot.scatter(x="A", y="B")\n-Out[101]: <matplotlib.collections.PathCollection at 0xe0f570e0>\n+Out[101]: <matplotlib.collections.PathCollection at 0xe09f10e0>\n
The hue
kwarg lets you vary the color by variable value
In [102]: ds.plot.scatter(x="A", y="B", hue="w")\n-Out[102]: <matplotlib.collections.PathCollection at 0xe2153450>\n+Out[102]: <matplotlib.collections.PathCollection at 0xe0bcd450>\n
You can force a legend instead of a colorbar by setting add_legend=True, add_colorbar=False
.
In [103]: ds.plot.scatter(x="A", y="B", hue="w", add_legend=True, add_colorbar=False)\n-Out[103]: <matplotlib.collections.PathCollection at 0xe0d8bb30>\n+Out[103]: <matplotlib.collections.PathCollection at 0xe0829b30>\n
In [104]: ds.plot.scatter(x="A", y="B", hue="w", add_legend=False, add_colorbar=True)\n-Out[104]: <matplotlib.collections.PathCollection at 0xe0d31190>\n+Out[104]: <matplotlib.collections.PathCollection at 0xe07cc190>\n
The markersize
kwarg lets you vary the point\u2019s size by variable value.\n You can additionally pass size_norm
to control how the variable\u2019s values are mapped to point sizes.
In [105]: ds.plot.scatter(x="A", y="B", hue="y", markersize="z")\n-Out[105]: <matplotlib.collections.PathCollection at 0xe0f9ff50>\n+Out[105]: <matplotlib.collections.PathCollection at 0xe0a3cf50>\n
The z
kwarg lets you plot the data along the z-axis as well.
In [106]: ds.plot.scatter(x="A", y="B", z="z", hue="y", markersize="x")\n-Out[106]: <mpl_toolkits.mplot3d.art3d.Path3DCollection at 0xe0cdf030>\n+Out[106]: <mpl_toolkits.mplot3d.art3d.Path3DCollection at 0xe077b030>\n
Faceting is also possible
\nIn [107]: ds.plot.scatter(x="A", y="B", hue="y", markersize="x", row="x", col="w")\n-Out[107]: <xarray.plot.facetgrid.FacetGrid at 0xe2402030>\n+Out[107]: <xarray.plot.facetgrid.FacetGrid at 0xe1ea3030>\n
And adding the z-axis
\nIn [108]: ds.plot.scatter(x="A", y="B", z="z", hue="y", markersize="x", row="x", col="w")\n-Out[108]: <xarray.plot.facetgrid.FacetGrid at 0xe0716df0>\n+Out[108]: <xarray.plot.facetgrid.FacetGrid at 0xe01b3df0>\n
For more advanced scatter plots, we recommend converting the relevant data variables\n to a pandas DataFrame and using the extensive plotting capabilities of seaborn
.
Visualizing vector fields is supported with quiver plots:
\nIn [109]: ds.isel(w=1, z=1).plot.quiver(x="x", y="y", u="A", v="B")\n-Out[109]: <matplotlib.quiver.Quiver at 0xe30ef760>\n+Out[109]: <matplotlib.quiver.Quiver at 0xe3339760>\n
where u
and v
denote the x and y direction components of the arrow vectors. Again, faceting is also possible:
In [110]: ds.plot.quiver(x="x", y="y", u="A", v="B", col="w", row="z", scale=4)\n-Out[110]: <xarray.plot.facetgrid.FacetGrid at 0xe3051f48>\n+Out[110]: <xarray.plot.facetgrid.FacetGrid at 0xe2b2ef48>\n
scale
is required for faceted quiver plots.\n The scale determines the number of data units per arrow length unit, i.e. a smaller scale parameter makes the arrow longer.
Visualizing vector fields is also supported with streamline plots:
\nIn [111]: ds.isel(w=1, z=1).plot.streamplot(x="x", y="y", u="A", v="B")\n-Out[111]: <matplotlib.collections.LineCollection at 0xe0aca3a0>\n+Out[111]: <matplotlib.collections.LineCollection at 0xe05673a0>\n
where u
and v
denote the x and y direction components of the vectors tangent to the streamlines.\n Again, faceting is also possible:
In [112]: ds.plot.streamplot(x="x", y="y", u="A", v="B", col="w", row="z")\n-Out[112]: <xarray.plot.facetgrid.FacetGrid at 0xe3071180>\n+Out[112]: <xarray.plot.facetgrid.FacetGrid at 0xe2b42180>\n
To follow this section you\u2019ll need to have Cartopy installed and working.
\nThis script will plot the air temperature on a map.
\nIn [113]: import cartopy.crs as ccrs\n \n In [114]: air = xr.tutorial.open_dataset("air_temperature").air\n-PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/first-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n+PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/second-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n \n \n In [115]: p = air.isel(time=0).plot(\n .....: subplot_kws=dict(projection=ccrs.Orthographic(-80, 35), facecolor="gray"),\n .....: transform=ccrs.PlateCarree(),\n .....: )\n .....: \n@@ -1024,24 +1024,24 @@\n In [121]: import xarray.plot as xplt\n \n In [122]: da = xr.DataArray(range(5))\n \n In [123]: fig, axs = plt.subplots(ncols=2, nrows=2)\n \n In [124]: da.plot(ax=axs[0, 0])\n-Out[124]: [<matplotlib.lines.Line2D at 0xdf3d6d40>]\n+Out[124]: [<matplotlib.lines.Line2D at 0xdee70d40>]\n \n In [125]: da.plot.line(ax=axs[0, 1])\n-Out[125]: [<matplotlib.lines.Line2D at 0xdf3d6f50>]\n+Out[125]: [<matplotlib.lines.Line2D at 0xdee70f50>]\n \n In [126]: xplt.plot(da, ax=axs[1, 0])\n-Out[126]: [<matplotlib.lines.Line2D at 0xdf3d6ea0>]\n+Out[126]: [<matplotlib.lines.Line2D at 0xdee70ea0>]\n \n In [127]: xplt.line(da, ax=axs[1, 1])\n-Out[127]: [<matplotlib.lines.Line2D at 0xdf3d6df0>]\n+Out[127]: [<matplotlib.lines.Line2D at 0xdee70df0>]\n \n In [128]: plt.tight_layout()\n \n In [129]: plt.draw()\n
\n \n
\n@@ -1091,15 +1091,15 @@\n
The plot will produce an image corresponding to the values of the array.\n Hence the top left pixel will be a different color than the others.\n Before reading on, you may want to look at the coordinates and\n think carefully about what the limits, labels, and orientation for\n each of the axes should be.
\nIn [134]: a.plot()\n-Out[134]: <matplotlib.collections.QuadMesh at 0xe09a5ea0>\n+Out[134]: <matplotlib.collections.QuadMesh at 0xe0443ea0>\n
It may seem strange that\n the values on the y axis are decreasing with -0.5 on the top. This is because\n the pixels are centered over their coordinates, and the\n@@ -1122,57 +1122,57 @@\n .....: np.arange(20).reshape(4, 5),\n .....: dims=["y", "x"],\n .....: coords={"lat": (("y", "x"), lat), "lon": (("y", "x"), lon)},\n .....: )\n .....: \n \n In [139]: da.plot.pcolormesh(x="lon", y="lat")\n-Out[139]: <matplotlib.collections.QuadMesh at 0xe00d20e0>\n+Out[139]: <matplotlib.collections.QuadMesh at 0xdfb700e0>\n
Note that in this case, xarray still follows the pixel centered convention.\n This might be undesirable in some cases, for example when your data is defined\n on a polar projection (GH781). This is why the default is to not follow\n this convention when plotting on a map:
\nIn [140]: import cartopy.crs as ccrs\n \n In [141]: ax = plt.subplot(projection=ccrs.PlateCarree())\n \n In [142]: da.plot.pcolormesh(x="lon", y="lat", ax=ax)\n-Out[142]: <cartopy.mpl.geocollection.GeoQuadMesh at 0xe0a01df0>\n+Out[142]: <cartopy.mpl.geocollection.GeoQuadMesh at 0xde79cdf0>\n \n In [143]: ax.scatter(lon, lat, transform=ccrs.PlateCarree())\n-Out[143]: <matplotlib.collections.PathCollection at 0xdf3de9d0>\n+Out[143]: <matplotlib.collections.PathCollection at 0xdee689d0>\n \n In [144]: ax.coastlines()\n-Out[144]: <cartopy.mpl.feature_artist.FeatureArtist at 0xe8528c68>\n+Out[144]: <cartopy.mpl.feature_artist.FeatureArtist at 0xea026c68>\n \n In [145]: ax.gridlines(draw_labels=True)\n-Out[145]: <cartopy.mpl.gridliner.Gridliner at 0xe8528818>\n+Out[145]: <cartopy.mpl.gridliner.Gridliner at 0xea026818>\n
You can however decide to infer the cell boundaries and use the\n infer_intervals
keyword:
In [146]: ax = plt.subplot(projection=ccrs.PlateCarree())\n \n In [147]: da.plot.pcolormesh(x="lon", y="lat", ax=ax, infer_intervals=True)\n-Out[147]: <cartopy.mpl.geocollection.GeoQuadMesh at 0xdecd3660>\n+Out[147]: <cartopy.mpl.geocollection.GeoQuadMesh at 0xde76d660>\n \n In [148]: ax.scatter(lon, lat, transform=ccrs.PlateCarree())\n-Out[148]: <matplotlib.collections.PathCollection at 0xdecdbd40>\n+Out[148]: <matplotlib.collections.PathCollection at 0xde773d40>\n \n In [149]: ax.coastlines()\n-Out[149]: <cartopy.mpl.feature_artist.FeatureArtist at 0xdecdbdf0>\n+Out[149]: <cartopy.mpl.feature_artist.FeatureArtist at 0xde773df0>\n \n In [150]: ax.gridlines(draw_labels=True)\n-Out[150]: <cartopy.mpl.gridliner.Gridliner at 0xdecdbea0>\n+Out[150]: <cartopy.mpl.gridliner.Gridliner at 0xde773ea0>\n
Note
\nThe data model of xarray does not support datasets with cell boundaries\n@@ -1180,26 +1180,26 @@\n outside the xarray framework.
\nOne can also make line plots with multidimensional coordinates. In this case, hue
must be a dimension name, not a coordinate name.
In [151]: f, ax = plt.subplots(2, 1)\n \n In [152]: da.plot.line(x="lon", hue="y", ax=ax[0])\n Out[152]: \n-[<matplotlib.lines.Line2D at 0xdecdb190>,\n- <matplotlib.lines.Line2D at 0xdeb59c90>,\n- <matplotlib.lines.Line2D at 0xdeb59d40>,\n- <matplotlib.lines.Line2D at 0xdeb59df0>]\n+[<matplotlib.lines.Line2D at 0xde773190>,\n+ <matplotlib.lines.Line2D at 0xde5f2c90>,\n+ <matplotlib.lines.Line2D at 0xde5f2d40>,\n+ <matplotlib.lines.Line2D at 0xde5f2df0>]\n \n In [153]: da.plot.line(x="lon", hue="x", ax=ax[1])\n Out[153]: \n-[<matplotlib.lines.Line2D at 0xdeb68190>,\n- <matplotlib.lines.Line2D at 0xdeb68240>,\n- <matplotlib.lines.Line2D at 0xdeb682f0>,\n- <matplotlib.lines.Line2D at 0xdeb683a0>,\n- <matplotlib.lines.Line2D at 0xdeb68450>]\n+[<matplotlib.lines.Line2D at 0xde602190>,\n+ <matplotlib.lines.Line2D at 0xde602240>,\n+ <matplotlib.lines.Line2D at 0xde6022f0>,\n+ <matplotlib.lines.Line2D at 0xde6023a0>,\n+ <matplotlib.lines.Line2D at 0xde602450>]\n
Whilst coarsen
is normally used for reducing your data\u2019s resolution by applying a reduction function\n (see the page on computation),\n it can also be used to reorganise your data without applying a computation via construct()
.
Taking our example tutorial air temperature dataset over the Northern US
\nIn [56]: air = xr.tutorial.open_dataset("air_temperature")["air"]\n-PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/first-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n+PermissionError: [Errno 13] Permission denied: '/nonexistent' | Pooch could not create data cache folder '/nonexistent/second-build/.cache/xarray_tutorial_data'. Will not be able to download data files.\n \n \n In [57]: air.isel(time=0).plot(x="lon", y="lat")\n NameError: name 'air' is not defined\n
To see an example of what each of these strategies might produce, you can call one followed by the .example()
method,\n which is a general hypothesis method valid for all strategies.
In [2]: import xarray.testing.strategies as xrst\n \n In [3]: xrst.variables().example()\n Out[3]: \n-<xarray.Variable (0: 1)> Size: 1B\n-array([0], dtype=int8)\n+<xarray.Variable (\u0123a: 3)> Size: 6B\n+array([-31411, -11635, -13617], dtype=int16)\n \n In [4]: xrst.variables().example()\n Out[4]: \n <xarray.Variable (0: 1)> Size: 1B\n array([0], dtype=int8)\n \n In [5]: xrst.variables().example()\n Out[5]: \n-<xarray.Variable (\u017e\u0175\u0116\u0147\u0140: 2, \u00bc: 3, \u00dd: 3)> Size: 72B\n-array([[[2613192455, 50, 12],\n- [ 19653, 4118, 2613192455],\n- [2613192455, 2613192455, 8022]],\n-\n- [[2613192455, 2613192455, 165],\n- [2613192455, 2613192455, 2613192455],\n- [2613192455, 2613192455, 2613192455]]], shape=(2, 3, 3), dtype=uint32)\n+<xarray.Variable (\u017c\u00b9\u017c\u0172: 4, 3\u013a: 2)> Size: 128B\n+array([[ nan-9.288e+295j, 3.648e+016-1.100e+000j],\n+ [ 8.888e+083+1.827e+016j, -1.500e+000-2.220e-016j],\n+ [ nan +infj, 1.000e+000+2.225e-311j],\n+ [-2.225e-309+1.100e+000j, inf +infj]])\n
You can see that calling .example()
multiple times will generate different examples, giving you an idea of the wide\n range of data that the xarray strategies can generate.
In your tests however you should not use .example()
- instead you should parameterize your tests with the\n hypothesis.given()
decorator:
In [6]: from hypothesis import given\n@@ -127,107 +124,115 @@\n Xarray\u2019s strategies can accept other strategies as arguments, allowing you to customise the contents of the generated\n examples.
\n # generate a Variable containing an array with a complex number dtype, but all other details still arbitrary\n In [8]: from hypothesis.extra.numpy import complex_number_dtypes\n \n In [9]: xrst.variables(dtype=complex_number_dtypes()).example()\n Out[9]: \n-<xarray.Variable (\u0124: 5)> Size: 80B\n-array([ inf-2.225e-308j, 1.000e-005-1.000e-005j, 3.106e-214+1.000e-005j,\n- 2.456e+016-1.000e-005j, 4.182e+016 +nanj], dtype='>c16')\n+<xarray.Variable (0: 1)> Size: 8B\n+array([0.+0.j], dtype='>c8')\n
\n \n This also works with custom strategies, or strategies defined in other packages.\n For example you could imagine creating a chunks
strategy to specify particular chunking patterns for a dask-backed array.
\n \n \n Fixing Arguments\u00b6
\n If you want to fix one aspect of the data structure, whilst allowing variation in the generated examples\n over all other aspects, then use hypothesis.strategies.just()
.
\n In [10]: import hypothesis.strategies as st\n \n # Generates only variable objects with dimensions ["x", "y"]\n In [11]: xrst.variables(dims=st.just(["x", "y"])).example()\n Out[11]: \n-<xarray.Variable (x: 4, y: 6)> Size: 192B\n-array([[ -52464, -786453851854018143, -52464, -52464,\n- -52464, -52464],\n- [ 10533, -52464, -52464, -52464,\n- -32324, -52464],\n- [-2917799381959081349, -3712367082676216948, -52464, 37242,\n- -52464, -1931],\n- [ 2460872617773150123, -3769050785953571223, -52464, -52464,\n- 7011038661760930734, -52464]], shape=(4, 6), dtype=int64)\n+<xarray.Variable (x: 6, y: 3)> Size: 144B\n+array([[29408, 29408, 4176],\n+ [29408, 29408, 29408],\n+ [29408, 15793, 29408],\n+ [29408, 53414, 29408],\n+ [29408, 29408, 29408],\n+ [29408, 29408, 29408]], shape=(6, 3), dtype=uint64)\n Attributes:\n- \u017c: {'\u017e': True, '\u00ef\u00cc\u017a\u010d': False, '': True, 'VM\u00be\u017f\u00c7': True}\n- 1: {'\u017e\u017f\u0122\u017c\u010c': array([nan+3.403e+38j]), 'm\u010b': None, '': False, '\u0127\u0146\u00dd'...\n- \u0143\u00e3\u00f9\u00d5: {'': False, '\u010f\u00f1\u011c\u00cf\u00f3': None}\n+ \u00e6tF\u00bc\u014a: {'\u0172': array([55157], dtype='>u8')}\n
\n \n (This is technically another example of chaining strategies - hypothesis.strategies.just()
is simply a\n special strategy that just contains a single example.)
\n To fix the length of dimensions you can instead pass dims
as a mapping of dimension names to lengths\n (i.e. following xarray objects\u2019 .sizes()
property), e.g.
\n # Generates only variables with dimensions ["x", "y"], of lengths 2 & 3 respectively\n In [12]: xrst.variables(dims=st.just({"x": 2, "y": 3})).example()\n Out[12]: \n-<xarray.Variable (x: 2, y: 3)> Size: 12B\n-array([[-0. , nan, -0. ],\n- [-1.9, 0. , -0. ]], dtype=float16)\n+<xarray.Variable (x: 2, y: 3)> Size: 24B\n+array([[ 363522030, 363522030, 234],\n+ [-1302831096, -980324081, 363522030]])\n+Attributes:\n+ \u0173\u00c2\u0106\u00fc\u00e2: {'\u017bH\u0151\u00daM': array([['\\xa0_\u00c8u\\U000d2a8a!\\x06\\x8a\u00b9\\U0008feb0\\x17', ...\n
\n \n You can also use this to specify that you want examples which are missing some part of the data structure, for instance
\n # Generates a Variable with no attributes\n In [13]: xrst.variables(attrs=st.just({})).example()\n Out[13]: \n-<xarray.Variable (\u017d\u015a\u017d3: 4, \u017e\u017e: 1, y\u017e: 3)> Size: 96B\n-array([[[-3.333e-001, -inf, 2.225e-309]],\n-\n- [[ nan, -2.225e-311, 3.333e-001]],\n-\n- [[ 6.104e-005, -1.175e-038, -2.949e+143]],\n-\n- [[ 1.113e-308, 6.400e+016, 9.007e+015]]], shape=(4, 1, 3))\n+<xarray.Variable (\u017c: 6, \u013e\u016a\u00ed\u00fd: 1)> Size: 12B\n+array([[35072],\n+ [14588],\n+ [19313],\n+ [14608],\n+ [37816],\n+ [19969]], dtype=uint16)\n
\n \n Through a combination of chaining strategies and fixing arguments, you can specify quite complicated requirements on the\n objects your chained strategy will generate.
\n In [14]: fixed_x_variable_y_maybe_z = st.fixed_dictionaries(\n ....: {"x": st.just(2), "y": st.integers(3, 4)}, optional={"z": st.just(2)}\n ....: )\n ....: \n \n In [15]: fixed_x_variable_y_maybe_z.example()\n-Out[15]: {'x': 2, 'y': 4}\n+Out[15]: {'x': 2, 'y': 3}\n \n In [16]: special_variables = xrst.variables(dims=fixed_x_variable_y_maybe_z)\n \n In [17]: special_variables.example()\n Out[17]: \n-<xarray.Variable (x: 2, y: 3, z: 2)> Size: 48B\n-array([[[ 100, 1227090441],\n- [ -24626, 32736],\n- [ -64858, 22392]],\n-\n- [[ -46726, -56716],\n- [ 51565, 51923],\n- [ 248, 13]]], shape=(2, 3, 2))\n+<xarray.Variable (x: 2, y: 4, z: 2)> Size: 128B\n+array([[[-1.000e+07 +nanj, 1.000e+00+7.043e+16j],\n+ [ 4.607e+16-0.000e+00j, 0.000e+00-3.378e+16j],\n+ [-1.000e+07 +nanj, -1.192e-07-2.220e-16j],\n+ [-5.000e-01-5.960e-08j, -1.000e+07 +nanj]],\n+\n+ [[ 0.000e+00 -infj, -0.000e+00-1.000e+00j],\n+ [ nan-6.213e+16j, 2.000e+00+9.007e+15j],\n+ [-5.960e-08 +nanj, -inf+0.000e+00j],\n+ [-6.312e+16 +nanj, -1.000e+07 +nanj]]], shape=(2, 4, 2), dtype=complex64)\n+Attributes:\n+ : True\n+ \u0167m\u017cCz: None\n+ \u00e9: ['RPi\\U0009dcdc']\n \n In [18]: special_variables.example()\n Out[18]: \n-<xarray.Variable (x: 2, y: 3, z: 2)> Size: 96B\n-array([[[ inf+1.000e+00j, 2.105e+16 +infj],\n- [-1.116e+16+1.900e+00j, 5.000e-01+1.900e+00j],\n- [-6.944e+16-5.002e+16j, -3.333e-01 +nanj]],\n-\n- [[ 1.000e-05 +infj, -1.192e-07-1.783e+16j],\n- [ 1.827e+16-2.000e+00j, -6.315e+16-0.000e+00j],\n- [ 1.000e-05 +infj, -inf+5.000e-01j]]], shape=(2, 3, 2), dtype=complex64)\n+<xarray.Variable (x: 2, y: 3, z: 2)> Size: 48B\n+array([[[-0.000e+00, nan],\n+ [ inf, 1.576e+16],\n+ [-6.081e+15, -inf]],\n+\n+ [[ 4.038e+16, 1.900e+00],\n+ [ 6.030e+16, 0.000e+00],\n+ [ inf, 0.000e+00]]], shape=(2, 3, 2), dtype=float32)\n Attributes:\n- \u017e\u00d2\u00f2\u00b3: None\n+ \u00bc\u0171\u0103: \u0173\n+ O: ['\u00d3']\n+ \u00ec: \u010f\u0100\u01133\n+ : [b'w\\xe3\\xcc\\xe8\\x1b\\x1a\\x9e\\xa8\\xf1']\n+ \u017d\u013d: \n+ \u00f8K\u0167\u017f\u0142: None\n+ \u00eb: [b'']\n
\n \n Here we have used one of hypothesis\u2019 built-in strategies hypothesis.strategies.fixed_dictionaries()
to create a\n strategy which generates mappings of dimension names to lengths (i.e. the size
of the xarray object we want).\n This particular strategy will always generate an x
dimension of length 2, and a y
dimension of\n length either 3 or 4, and will sometimes also generate a z
dimension of length 2.\n By feeding this strategy for dictionaries into the dims
argument of xarray\u2019s variables()
strategy,\n@@ -328,53 +333,43 @@\n ....: array_strategy_fn=xps.arrays,\n ....: dtype=xps.scalar_dtypes(),\n ....: )\n ....: \n \n In [32]: xp_variables.example()\n Out[32]: \n-<xarray.Variable (\u017d\u0136\u0139\u012c\u0102: 1)> Size: 1B\n-array([230], dtype=uint8)\n-Attributes:\n- \u00d0K\u012eq\u00cf: \u017b2\u017d\u017b\u017d\n- : True\n- R: [[ True False]\\n [False True]]\n- 5\u00e5: None\n- \u0121\u00fd: ['\\x88\\x1a\\U00106ee4\\x01H\ud872\udcbf6' '\\x96\\U00073224\u00d9\\x04(\u00e6\\U000fbd18\\x...\n- \u0167\u017b\u014c\u0151\u0101: \u012bM\u0165\u017ea\n- \u017b: None\n- \u0104\u013c\u00e2\u017e8: [34367 21905]\n- 5\u0127: None\n+<xarray.Variable (\u00ec\u017c: 3)> Size: 24B\n+array([ inf-0.000e+00j, -0.000e+00-3.403e+38j, 3.403e+38 -infj], dtype=complex64)\n
Another array API-compliant duck array library would replace the import, e.g. import cupy as cp
instead.
A common task when testing xarray user code is checking that your function works for all valid input dimensions.\n We can chain strategies to achieve this, for which the helper strategy unique_subset_of()
\n is useful.
It works for lists of dimension names
\nIn [33]: dims = ["x", "y", "z"]\n \n In [34]: xrst.unique_subset_of(dims).example()\n-Out[34]: ['z', 'y', 'x']\n+Out[34]: ['y', 'z']\n \n In [35]: xrst.unique_subset_of(dims).example()\n-Out[35]: ['y', 'z', 'x']\n+Out[35]: ['z']\n
as well as for mappings of dimension names to sizes
\nIn [36]: dim_sizes = {"x": 2, "y": 3, "z": 4}\n \n In [37]: xrst.unique_subset_of(dim_sizes).example()\n-Out[37]: {'y': 3, 'z': 4}\n+Out[37]: {'x': 2, 'y': 3, 'z': 4}\n \n In [38]: xrst.unique_subset_of(dim_sizes).example()\n-Out[38]: {'z': 4, 'x': 2, 'y': 3}\n+Out[38]: {'x': 2}\n
This is useful because operations like reductions can be performed over any subset of the xarray object\u2019s dimensions.\n For example we can write a pytest test that tests that a reduction gives the expected result when applying that reduction\n along any possible valid subset of the Variable\u2019s dimensions.
\nimport numpy.testing as npt\n \n", "details": [{"source1": "html2text {}", "source2": "html2text {}", "unified_diff": "@@ -28,32 +28,29 @@\n To see an example of what each of these strategies might produce, you can call\n one followed by the .example() method, which is a general hypothesis method\n valid for all strategies.\n In [2]: import xarray.testing.strategies as xrst\n \n In [3]: xrst.variables().example()\n Out[3]:\n- Size: 1B\n-array([0], dtype=int8)\n+ Size: 6B\n+array([-31411, -11635, -13617], dtype=int16)\n \n In [4]: xrst.variables().example()\n Out[4]:\n Size: 1B\n array([0], dtype=int8)\n \n In [5]: xrst.variables().example()\n Out[5]:\n- Size: 72B\n-array([[[2613192455, 50, 12],\n- [ 19653, 4118, 2613192455],\n- [2613192455, 2613192455, 8022]],\n-\n- [[2613192455, 2613192455, 165],\n- [2613192455, 2613192455, 2613192455],\n- [2613192455, 2613192455, 2613192455]]], shape=(2, 3, 3), dtype=uint32)\n+ Size: 128B\n+array([[ nan-9.288e+295j, 3.648e+016-1.100e+000j],\n+ [ 8.888e+083+1.827e+016j, -1.500e+000-2.220e-016j],\n+ [ nan +infj, 1.000e+000+2.225e-311j],\n+ [-2.225e-309+1.100e+000j, inf +infj]])\n You can see that calling .example() multiple times will generate different\n examples, giving you an idea of the wide range of data that the xarray\n strategies can generate.\n In your tests however you should not use .example() - instead you should\n parameterize your tests with the hypothesis.given() decorator:\n In [6]: from hypothesis import given\n In [7]: @given(xrst.variables())\n@@ -65,112 +62,115 @@\n customise the contents of the generated examples.\n # generate a Variable containing an array with a complex number dtype, but all\n other details still arbitrary\n In [8]: from hypothesis.extra.numpy import complex_number_dtypes\n \n In [9]: xrst.variables(dtype=complex_number_dtypes()).example()\n Out[9]:\n- Size: 80B\n-array([ inf-2.225e-308j, 1.000e-005-1.000e-005j, 3.106e-214+1.000e-005j,\n- 2.456e+016-1.000e-005j, 4.182e+016 +nanj], dtype='>c16')\n+ Size: 8B\n+array([0.+0.j], dtype='>c8')\n This also works with custom strategies, or strategies defined in other\n packages. For example you could imagine creating a chunks strategy to specify\n particular chunking patterns for a dask-backed array.\n *\b**\b**\b**\b* F\bFi\bix\bxi\bin\bng\bg A\bAr\brg\bgu\bum\bme\ben\bnt\bts\bs_\b?\b\u00b6 *\b**\b**\b**\b*\n If you want to fix one aspect of the data structure, whilst allowing variation\n in the generated examples over all other aspects, then use\n hypothesis.strategies.just().\n In [10]: import hypothesis.strategies as st\n \n # Generates only variable objects with dimensions [\"x\", \"y\"]\n In [11]: xrst.variables(dims=st.just([\"x\", \"y\"])).example()\n Out[11]:\n- Size: 192B\n-array([[ -52464, -786453851854018143, -52464,\n--52464,\n- -52464, -52464],\n- [ 10533, -52464, -52464,\n--52464,\n- -32324, -52464],\n- [-2917799381959081349, -3712367082676216948, -52464,\n-37242,\n- -52464, -1931],\n- [ 2460872617773150123, -3769050785953571223, -52464,\n--52464,\n- 7011038661760930734, -52464]], shape=(4, 6),\n-dtype=int64)\n+ Size: 144B\n+array([[29408, 29408, 4176],\n+ [29408, 29408, 29408],\n+ [29408, 15793, 29408],\n+ [29408, 53414, 29408],\n+ [29408, 29408, 29408],\n+ [29408, 29408, 29408]], shape=(6, 3), dtype=uint64)\n Attributes:\n- \u017c: {'\u017e': True, '\u00ef\u00cc\u017a\u010d': False, '': True, 'VM\u00be\u017f\u00c7': True}\n- 1: {'\u017e\u017f\u0122\u017c\u010c': array([nan+3.403e+38j]), 'm\u010b': None, '': False,\n-'\u0127\u0146\u00dd'...\n- \u0143\u00e3\u00f9\u00d5: {'': False, '\u010f\u00f1\u011c\u00cf\u00f3': None}\n+ \u00e6tF\u00bc\u014a: {'\u0172': array([55157], dtype='>u8')}\n (This is technically another example of chaining strategies -\n hypothesis.strategies.just() is simply a special strategy that just contains a\n single example.)\n To fix the length of dimensions you can instead pass dims as a mapping of\n dimension names to lengths (i.e. following xarray objects\u2019 .sizes() property),\n e.g.\n # Generates only variables with dimensions [\"x\", \"y\"], of lengths 2 & 3\n respectively\n In [12]: xrst.variables(dims=st.just({\"x\": 2, \"y\": 3})).example()\n Out[12]:\n- Size: 12B\n-array([[-0. , nan, -0. ],\n- [-1.9, 0. , -0. ]], dtype=float16)\n+ Size: 24B\n+array([[ 363522030, 363522030, 234],\n+ [-1302831096, -980324081, 363522030]])\n+Attributes:\n+ \u0173\u00c2\u0106\u00fc\u00e2: {'\u017bH\u0151\u00daM': array([['\\xa0_\u00c8u\\U000d2a8a!\\x06\\x8a\u00b9\\U0008feb0\\x17',\n+...\n You can also use this to specify that you want examples which are missing some\n part of the data structure, for instance\n # Generates a Variable with no attributes\n In [13]: xrst.variables(attrs=st.just({})).example()\n Out[13]:\n- Size: 96B\n-array([[[-3.333e-001, -inf, 2.225e-309]],\n-\n- [[ nan, -2.225e-311, 3.333e-001]],\n-\n- [[ 6.104e-005, -1.175e-038, -2.949e+143]],\n-\n- [[ 1.113e-308, 6.400e+016, 9.007e+015]]], shape=(4, 1, 3))\n+ Size: 12B\n+array([[35072],\n+ [14588],\n+ [19313],\n+ [14608],\n+ [37816],\n+ [19969]], dtype=uint16)\n Through a combination of chaining strategies and fixing arguments, you can\n specify quite complicated requirements on the objects your chained strategy\n will generate.\n In [14]: fixed_x_variable_y_maybe_z = st.fixed_dictionaries(\n ....: {\"x\": st.just(2), \"y\": st.integers(3, 4)}, optional={\"z\": st.just\n (2)}\n ....: )\n ....:\n \n In [15]: fixed_x_variable_y_maybe_z.example()\n-Out[15]: {'x': 2, 'y': 4}\n+Out[15]: {'x': 2, 'y': 3}\n \n In [16]: special_variables = xrst.variables(dims=fixed_x_variable_y_maybe_z)\n \n In [17]: special_variables.example()\n Out[17]:\n- Size: 48B\n-array([[[ 100, 1227090441],\n- [ -24626, 32736],\n- [ -64858, 22392]],\n-\n- [[ -46726, -56716],\n- [ 51565, 51923],\n- [ 248, 13]]], shape=(2, 3, 2))\n+ Size: 128B\n+array([[[-1.000e+07 +nanj, 1.000e+00+7.043e+16j],\n+ [ 4.607e+16-0.000e+00j, 0.000e+00-3.378e+16j],\n+ [-1.000e+07 +nanj, -1.192e-07-2.220e-16j],\n+ [-5.000e-01-5.960e-08j, -1.000e+07 +nanj]],\n+\n+ [[ 0.000e+00 -infj, -0.000e+00-1.000e+00j],\n+ [ nan-6.213e+16j, 2.000e+00+9.007e+15j],\n+ [-5.960e-08 +nanj, -inf+0.000e+00j],\n+ [-6.312e+16 +nanj, -1.000e+07 +nanj]]], shape=(2, 4, 2),\n+dtype=complex64)\n+Attributes:\n+ : True\n+ \u0167m\u017cCz: None\n+ \u00e9: ['RPi\\U0009dcdc']\n \n In [18]: special_variables.example()\n Out[18]:\n- Size: 96B\n-array([[[ inf+1.000e+00j, 2.105e+16 +infj],\n- [-1.116e+16+1.900e+00j, 5.000e-01+1.900e+00j],\n- [-6.944e+16-5.002e+16j, -3.333e-01 +nanj]],\n-\n- [[ 1.000e-05 +infj, -1.192e-07-1.783e+16j],\n- [ 1.827e+16-2.000e+00j, -6.315e+16-0.000e+00j],\n- [ 1.000e-05 +infj, -inf+5.000e-01j]]], shape=(2, 3, 2),\n-dtype=complex64)\n+ Size: 48B\n+array([[[-0.000e+00, nan],\n+ [ inf, 1.576e+16],\n+ [-6.081e+15, -inf]],\n+\n+ [[ 4.038e+16, 1.900e+00],\n+ [ 6.030e+16, 0.000e+00],\n+ [ inf, 0.000e+00]]], shape=(2, 3, 2), dtype=float32)\n Attributes:\n- \u017e\u00d2\u00f2\u00b3: None\n+ \u00bc\u0171\u0103: \u0173\n+ O: ['\u00d3']\n+ \u00ec: \u010f\u0100\u01133\n+ : [b'w\\xe3\\xcc\\xe8\\x1b\\x1a\\x9e\\xa8\\xf1']\n+ \u017d\u013d:\n+ \u00f8K\u0167\u017f\u0142: None\n+ \u00eb: [b'']\n Here we have used one of hypothesis\u2019 built-in strategies\n hypothesis.strategies.fixed_dictionaries() to create a strategy which generates\n mappings of dimension names to lengths (i.e. the size of the xarray object we\n want). This particular strategy will always generate an x dimension of length\n 2, and a y dimension of length either 3 or 4, and will sometimes also generate\n a z dimension of length 2. By feeding this strategy for dictionaries into the\n dims argument of xarray\u2019s variables() strategy, we can generate arbitrary\n@@ -264,49 +264,39 @@\n ....: array_strategy_fn=xps.arrays,\n ....: dtype=xps.scalar_dtypes(),\n ....: )\n ....:\n \n In [32]: xp_variables.example()\n Out[32]:\n- Size: 1B\n-array([230], dtype=uint8)\n-Attributes:\n- \u00d0K\u012eq\u00cf: \u017b2\u017d\u017b\u017d\n- : True\n- R: [[ True False]\\n [False True]]\n- 5\u00e5: None\n- \u0121\u00fd: ['\\x88\\x1a\\U00106ee4\\x01H\ud872\udcbf6' '\\x96\\U00073224\u00d9\\x04\n-(\u00e6\\U000fbd18\\x...\n- \u0167\u017b\u014c\u0151\u0101: \u012bM\u0165\u017ea\n- \u017b: None\n- \u0104\u013c\u00e2\u017e8: [34367 21905]\n- 5\u0127: None\n+ Size: 24B\n+array([ inf-0.000e+00j, -0.000e+00-3.403e+38j, 3.403e+38 -infj],\n+dtype=complex64)\n Another array API-compliant duck array library would replace the import, e.g.\n import cupy as cp instead.\n *\b**\b**\b**\b* T\bTe\bes\bst\bti\bin\bng\bg o\bov\bve\ber\br S\bSu\bub\bbs\bse\bet\bts\bs o\bof\bf D\bDi\bim\bme\ben\bns\bsi\bio\bon\bns\bs_\b?\b\u00b6 *\b**\b**\b**\b*\n A common task when testing xarray user code is checking that your function\n works for all valid input dimensions. We can chain strategies to achieve this,\n for which the helper strategy unique_subset_of() is useful.\n It works for lists of dimension names\n In [33]: dims = [\"x\", \"y\", \"z\"]\n \n In [34]: xrst.unique_subset_of(dims).example()\n-Out[34]: ['z', 'y', 'x']\n+Out[34]: ['y', 'z']\n \n In [35]: xrst.unique_subset_of(dims).example()\n-Out[35]: ['y', 'z', 'x']\n+Out[35]: ['z']\n as well as for mappings of dimension names to sizes\n In [36]: dim_sizes = {\"x\": 2, \"y\": 3, \"z\": 4}\n \n In [37]: xrst.unique_subset_of(dim_sizes).example()\n-Out[37]: {'y': 3, 'z': 4}\n+Out[37]: {'x': 2, 'y': 3, 'z': 4}\n \n In [38]: xrst.unique_subset_of(dim_sizes).example()\n-Out[38]: {'z': 4, 'x': 2, 'y': 3}\n+Out[38]: {'x': 2}\n This is useful because operations like reductions can be performed over any\n subset of the xarray object\u2019s dimensions. For example we can write a pytest\n test that tests that a reduction gives the expected result when applying that\n reduction along any possible valid subset of the Variable\u2019s dimensions.\n import numpy.testing as npt\n \n \n"}]}, {"source1": "./usr/share/doc/python-xarray-doc/html/whats-new.html", "source2": "./usr/share/doc/python-xarray-doc/html/whats-new.html", "unified_diff": "@@ -7936,15 +7936,15 @@\n New xray.Dataset.where
method for masking xray objects according\n to some criteria. This works particularly well with multi-dimensional data:
\n In [44]: ds = xray.Dataset(coords={"x": range(100), "y": range(100)})\n \n In [45]: ds["distance"] = np.sqrt(ds.x**2 + ds.y**2)\n \n In [46]: ds.distance.where(ds.distance < 100).plot()\n-Out[46]: <matplotlib.collections.QuadMesh at 0xde6e17c0>\n+Out[46]: <matplotlib.collections.QuadMesh at 0xde07abe0>\n
\n \n
\n \n \n Added new methods xray.DataArray.diff
and xray.Dataset.diff
\n for finite difference calculations along a given axis.
\n", "details": [{"source1": "html2text {}", "source2": "html2text {}", "unified_diff": "@@ -5100,15 +5100,15 @@\n * New xray.Dataset.where method for masking xray objects according to some\n criteria. This works particularly well with multi-dimensional data:\n In [44]: ds = xray.Dataset(coords={\"x\": range(100), \"y\": range(100)})\n \n In [45]: ds[\"distance\"] = np.sqrt(ds.x**2 + ds.y**2)\n \n In [46]: ds.distance.where(ds.distance < 100).plot()\n- Out[46]: \n+ Out[46]: \n _\b[_\b__\bb_\bu_\bi_\bl_\bd_\b/_\bh_\bt_\bm_\bl_\b/_\b__\bs_\bt_\ba_\bt_\bi_\bc_\b/_\bw_\bh_\be_\br_\be_\b__\be_\bx_\ba_\bm_\bp_\bl_\be_\b._\bp_\bn_\bg_\b]\n * Added new methods xray.DataArray.diff and xray.Dataset.diff for finite\n difference calculations along a given axis.\n * New xray.DataArray.to_masked_array convenience method for returning a\n numpy.ma.MaskedArray.\n In [47]: da = xray.DataArray(np.random.random_sample(size=(5, 4)))\n \n"}]}]}]}]}]}