Skip to content

Commit

Permalink
Purge dask-expr
Browse files Browse the repository at this point in the history
  • Loading branch information
hendrikmakait committed Jan 9, 2025
1 parent 615ba99 commit 77911ec
Show file tree
Hide file tree
Showing 9 changed files with 7 additions and 14 deletions.
2 changes: 0 additions & 2 deletions AB_environments/AB_sample.conda.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,6 @@ dependencies:
- pandas ==2.2.3
- dask ==2024.11.2
- distributed ==2024.11.2
- dask-expr ==1.1.19
- dask-labextension ==7.0.0
- dask-ml ==2024.4.4
- fsspec ==2024.10.0
Expand Down Expand Up @@ -74,4 +73,3 @@ dependencies:
# Read README.md for troubleshooting.
- git+https://github.com/dask/dask@191d39177009d2cce25b818878118e35329b6db3
- git+https://github.com/dask/distributed@0304fb6e665e36abf9e3086173cccd36e29ae84d
- git+https://github.com/dask-contrib/dask-expr@9f765764da3f518ddd4c896c98b8a40a979a5553
1 change: 0 additions & 1 deletion AB_environments/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -222,7 +222,6 @@ As a handy copy-paste to run from the root dir of this repository:
```bash
pushd ../dask && git fetch upstream --tags && git push origin --tags && popd
pushd ../distributed && git fetch upstream --tags && git push origin --tags && popd
pushd ../dask-expr && git fetch upstream --tags && git push origin --tags && popd
```

#### Problem:
Expand Down
1 change: 0 additions & 1 deletion ci/environment-dashboard.yml
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,6 @@ dependencies:
- coiled
- conda
- dask
- dask-expr
- dask-ml
- distributed
- filelock
Expand Down
6 changes: 3 additions & 3 deletions tests/benchmarks/test_join.py
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ def test_join_big(small_client, memory_multiplier):
df2_big = df2_big.astype({"predicate": "int"})

join = df1_big.merge(df2_big, on="predicate", how="inner")
# dask-expr will drop all columns except the Index for size
# dask.dataframe will drop all columns except the Index for size
# computations, which will optimize itself through merges, e.g.
# shuffling a lot less data than what we want to test
# map_partitions blocks those optimizations
Expand Down Expand Up @@ -55,7 +55,7 @@ def test_join_big_small(small_client, memory_multiplier, configure_shuffling):
df_small_pd = df_small.astype({"predicate": "int"}).compute()

join = df_big.merge(df_small_pd, on="predicate", how="inner")
# dask-expr will drop all columns except the Index for size
# dask.dataframe will drop all columns except the Index for size
# computations, which will optimize itself through merges, e.g.
# shuffling a lot less data than what we want to test
# map_partitions blocks those optimizations
Expand All @@ -77,7 +77,7 @@ def test_set_index(small_client, persist, memory_multiplier):
if persist:
df_big = df_big.persist()
df_indexed = df_big.set_index("0")
# dask-expr will drop all columns except the Index for size
# dask.dataframe will drop all columns except the Index for size
# computations, which will optimize itself through set_index, e.g.
# shuffling a lot less data than what we want to test
# map_partitions blocks those optimizations
Expand Down
2 changes: 0 additions & 2 deletions tests/conftest.py
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,6 @@
import adlfs
import dask
import dask.array as da
import dask_expr
import distributed
import filelock
import gcsfs
Expand Down Expand Up @@ -174,7 +173,6 @@ def database_table_schema(request, testrun_uid):
originalname=request.node.originalname,
path=str(request.node.path.relative_to(TEST_DIR)),
dask_version=dask.__version__,
dask_expr_version=dask_expr.__version__,
distributed_version=distributed.__version__,
coiled_runtime_version=os.environ.get("AB_VERSION", "upstream"),
coiled_software_name=COILED_SOFTWARE_NAME,
Expand Down
3 changes: 2 additions & 1 deletion tests/geospatial/test_climatology.py
Original file line number Diff line number Diff line change
Expand Up @@ -47,10 +47,11 @@ def test_highlevel_api(
"workspace": "dask-benchmarks-gcp",
"region": "us-central1",
"idle_timeout": "1h",
"scheduler_memory": "64 GiB",
},
scale_kwargs={
"small": {"n_workers": 10},
"medium": {"n_workers": 100},
"medium": {"n_workers": 10},
"large": {"n_workers": 100},
},
):
Expand Down
2 changes: 0 additions & 2 deletions tests/tpch/conftest.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,6 @@

import coiled
import dask
import dask_expr
import distributed
import filelock
import pytest
Expand Down Expand Up @@ -142,7 +141,6 @@ def tpch_database_table_schema(request, testrun_uid, scale, query, local):
originalname=request.node.originalname,
path=str(request.node.path.relative_to(TEST_DIR)),
dask_version=dask.__version__,
dask_expr_version=dask_expr.__version__,
distributed_version=distributed.__version__,
python_version=".".join(map(str, sys.version_info)),
platform=sys.platform,
Expand Down
2 changes: 1 addition & 1 deletion tests/tpch/dask_queries.py
Original file line number Diff line number Diff line change
Expand Up @@ -949,7 +949,7 @@ def query_18(dataset_path, fs, scale):
orders = dd.read_parquet(dataset_path + "orders", filesystem=fs)
lineitem = dd.read_parquet(dataset_path + "lineitem", filesystem=fs)

# FIXME: https://github.com/dask-contrib/dask-expr/issues/867
# FIXME: https://github.com/dask/dask-expr/issues/867
qnt_over_300 = (
lineitem.groupby("l_orderkey").l_quantity.sum(split_out=True).reset_index()
)
Expand Down
2 changes: 1 addition & 1 deletion tests/tpch/test_dask.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@
@pytest.fixture(scope="session")
def dataset_path(local, scale):
if local:
# FIXME: pyarrow local fs is a bit odd. dask-expr should deal with this
# FIXME: pyarrow local fs is a bit odd. dask.dataframe should deal with this
return "file://" + os.path.abspath(get_dataset_path(local, scale)) + "/"
else:
return get_dataset_path(local, scale)
Expand Down

0 comments on commit 77911ec

Please sign in to comment.