Skip to content

Commit

Permalink
mark tests as needing pyarrow
Browse files Browse the repository at this point in the history
  • Loading branch information
graingert committed Dec 7, 2022
1 parent 08fd458 commit e809ae1
Showing 1 changed file with 3 additions and 0 deletions.
3 changes: 3 additions & 0 deletions dask/dataframe/io/tests/test_parquet.py
Original file line number Diff line number Diff line change
Expand Up @@ -661,6 +661,7 @@ def write_partition(df, i):
assert_eq(df, ddf2, check_index=False)


@PYARROW_MARK
@pytest.mark.xfail(
not PANDAS_GT_130,
reason=(
Expand Down Expand Up @@ -3005,6 +3006,7 @@ def test_chunksize_aggregate_files(tmpdir, write_engine, read_engine, aggregate_
assert_eq(df1[["c", "d"]], df2[["c", "d"]], check_index=False)


@PYARROW_MARK
@pytest.mark.parametrize("metadata", [True, False])
@pytest.mark.parametrize("chunksize", [None, 1024, 4096, "1MiB"])
def test_chunksize(tmpdir, chunksize, engine, metadata):
Expand Down Expand Up @@ -3998,6 +4000,7 @@ def test_metadata_task_size(tmpdir, engine, write_metadata_file, metadata_task_s
assert_eq(ddf2b, ddf2c)


@PYARROW_MARK
@pytest.mark.parametrize("partition_on", ("b", None))
def test_extra_file(tmpdir, engine, partition_on):
# Check that read_parquet can handle spark output
Expand Down

0 comments on commit e809ae1

Please sign in to comment.