Skip to content

Commit

Permalink
Avoid set_index(..., inplace=True) where not necessary (#9472)
Browse files Browse the repository at this point in the history
  • Loading branch information
jrbourbeau committed Sep 15, 2022
1 parent f45df2b commit 803c7fd
Show file tree
Hide file tree
Showing 3 changed files with 6 additions and 5 deletions.
1 change: 1 addition & 0 deletions dask/dataframe/io/orc/core.py
Expand Up @@ -49,6 +49,7 @@ def __call__(self, parts):
)
if self.index:
_df.set_index(self.index, inplace=True)

return _df


Expand Down
2 changes: 1 addition & 1 deletion dask/dataframe/io/tests/test_orc.py
Expand Up @@ -97,7 +97,7 @@ def test_orc_roundtrip(tmpdir, index, columns):
}
)
if index:
data.set_index(index, inplace=True)
data = data.set_index(index)
df = dd.from_pandas(data, chunksize=500)
if columns:
data = data[[c for c in columns if c != index]]
Expand Down
8 changes: 4 additions & 4 deletions dask/dataframe/io/tests/test_parquet.py
Expand Up @@ -196,7 +196,7 @@ def test_empty(tmpdir, write_engine, read_engine, index):
fn = str(tmpdir)
df = pd.DataFrame({"a": ["a", "b", "b"], "b": [4, 5, 6]})[:0]
if index:
df.set_index("a", inplace=True, drop=True)
df = df.set_index("a", drop=True)
ddf = dd.from_pandas(df, npartitions=2)

ddf.to_parquet(fn, write_index=index, engine=write_engine, write_metadata_file=True)
Expand All @@ -208,7 +208,7 @@ def test_empty(tmpdir, write_engine, read_engine, index):
def test_simple(tmpdir, write_engine, read_engine):
fn = str(tmpdir)
df = pd.DataFrame({"a": ["a", "b", "b"], "b": [4, 5, 6]})
df.set_index("a", inplace=True, drop=True)
df = df.set_index("a", drop=True)
ddf = dd.from_pandas(df, npartitions=2)
ddf.to_parquet(fn, engine=write_engine)
read_df = dd.read_parquet(
Expand All @@ -221,7 +221,7 @@ def test_simple(tmpdir, write_engine, read_engine):
def test_delayed_no_metadata(tmpdir, write_engine, read_engine):
fn = str(tmpdir)
df = pd.DataFrame({"a": ["a", "b", "b"], "b": [4, 5, 6]})
df.set_index("a", inplace=True, drop=True)
df = df.set_index("a", drop=True)
ddf = dd.from_pandas(df, npartitions=2)
ddf.to_parquet(
fn, engine=write_engine, compute=False, write_metadata_file=False
Expand Down Expand Up @@ -3333,7 +3333,7 @@ def test_divisions_with_null_partition(tmpdir, engine):
def test_pyarrow_dataset_simple(tmpdir, engine):
fn = str(tmpdir)
df = pd.DataFrame({"a": [4, 5, 6], "b": ["a", "b", "b"]})
df.set_index("a", inplace=True, drop=True)
df = df.set_index("a", drop=True)
ddf = dd.from_pandas(df, npartitions=2)
ddf.to_parquet(fn, engine=engine)
read_df = dd.read_parquet(fn, engine="pyarrow", calculate_divisions=True)
Expand Down

0 comments on commit 803c7fd

Please sign in to comment.