File tree 2 files changed +5
-2
lines changed
2 files changed +5
-2
lines changed Original file line number Diff line number Diff line change @@ -61,7 +61,7 @@ Internal Changes
61
61
- Use Python 3.6 idioms throughout the codebase. (:pull: 3419)
62
62
By `Maximilian Roos <https://github.com/max-sixty >`_
63
63
- Implement :py:func: `__dask_tokenize__ ` for xarray objects.
64
- By `Deepak Cherian <https://github.com/dcherian >`_
64
+ By `Deepak Cherian <https://github.com/dcherian >`_ and ` Guido Imperiale < https://github.com/crusaderky >`_.
65
65
66
66
.. _whats-new.0.14.0 :
67
67
Original file line number Diff line number Diff line change 1
1
import operator
2
2
import pickle
3
+ import sys
3
4
from contextlib import suppress
4
5
from distutils .version import LooseVersion
5
6
from textwrap import dedent
28
29
da = pytest .importorskip ("dask.array" )
29
30
dd = pytest .importorskip ("dask.dataframe" )
30
31
32
+ ON_WINDOWS = sys .platform == "win32"
33
+
31
34
32
35
class CountingScheduler :
33
36
""" Simple dask scheduler counting the number of computes.
@@ -1186,7 +1189,7 @@ def test_normalize_token_identical(obj, transform):
1186
1189
1187
1190
1188
1191
def test_normalize_token_netcdf_backend (map_ds ):
1189
- with create_tmp_file () as tmp_file :
1192
+ with create_tmp_file (allow_cleanup_failure = ON_WINDOWS ) as tmp_file :
1190
1193
map_ds .to_netcdf (tmp_file )
1191
1194
read = xr .open_dataset (tmp_file )
1192
1195
assert not dask .base .tokenize (map_ds ) == dask .base .tokenize (read )
You can’t perform that action at this time.
0 commit comments