Skip to content

Commit

Permalink
CI: update fastparquet xfails for new release (pandas-dev#60337)
Browse files Browse the repository at this point in the history
(cherry picked from commit 34c080c)
  • Loading branch information
WillAyd authored and jorisvandenbossche committed Nov 17, 2024
1 parent 7958d6c commit c365dfc
Show file tree
Hide file tree
Showing 3 changed files with 39 additions and 16 deletions.
6 changes: 5 additions & 1 deletion pandas/tests/io/test_fsspec.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,8 @@

from pandas._config import using_string_dtype

from pandas.compat import HAS_PYARROW

from pandas import (
DataFrame,
date_range,
Expand Down Expand Up @@ -168,7 +170,9 @@ def test_excel_options(fsspectest):
assert fsspectest.test[0] == "read"


@pytest.mark.xfail(using_string_dtype(), reason="TODO(infer_string) fastparquet")
@pytest.mark.xfail(
using_string_dtype() and HAS_PYARROW, reason="TODO(infer_string) fastparquet"
)
def test_to_parquet_new_file(cleared_fs, df1):
"""Regression test for writing to a not-yet-existent GCS Parquet file."""
pytest.importorskip("fastparquet")
Expand Down
3 changes: 0 additions & 3 deletions pandas/tests/io/test_gcs.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@
import numpy as np
import pytest

from pandas._config import using_string_dtype

from pandas.compat.pyarrow import pa_version_under17p0

from pandas import (
Expand Down Expand Up @@ -196,7 +194,6 @@ def test_to_csv_compression_encoding_gcs(
tm.assert_frame_equal(df, read_df)


@pytest.mark.xfail(using_string_dtype(), reason="TODO(infer_string) fastparquet")
def test_to_parquet_gcs_new_file(monkeypatch, tmpdir):
"""Regression test for writing to a not-yet-existent GCS Parquet file."""
pytest.importorskip("fastparquet")
Expand Down
46 changes: 34 additions & 12 deletions pandas/tests/io/test_parquet.py
Original file line number Diff line number Diff line change
Expand Up @@ -1187,7 +1187,16 @@ def test_infer_string_large_string_type(self, tmp_path, pa):


class TestParquetFastParquet(Base):
def test_basic(self, fp, df_full):
def test_basic(self, fp, df_full, request):
import fastparquet

if Version(fastparquet.__version__) < Version("2024.11.0"):
request.applymarker(
pytest.mark.xfail(
reason=("datetime_with_nat gets incorrect values"),
)
)

df = df_full

dti = pd.date_range("20130101", periods=3, tz="US/Eastern")
Expand Down Expand Up @@ -1223,11 +1232,17 @@ def test_duplicate_columns(self, fp):
msg = "Cannot create parquet dataset with duplicate column names"
self.check_error_on_write(df, fp, ValueError, msg)

@pytest.mark.xfail(
Version(np.__version__) >= Version("2.0.0"),
reason="fastparquet uses np.float_ in numpy2",
)
def test_bool_with_none(self, fp):
def test_bool_with_none(self, fp, request):
import fastparquet

if Version(fastparquet.__version__) < Version("2024.11.0") and Version(
np.__version__
) >= Version("2.0.0"):
request.applymarker(
pytest.mark.xfail(
reason=("fastparquet uses np.float_ in numpy2"),
)
)
df = pd.DataFrame({"a": [True, None, False]})
expected = pd.DataFrame({"a": [1.0, np.nan, 0.0]}, dtype="float16")
# Fastparquet bug in 0.7.1 makes it so that this dtype becomes
Expand Down Expand Up @@ -1342,12 +1357,19 @@ def test_empty_dataframe(self, fp):
expected = df.copy()
check_round_trip(df, fp, expected=expected)

@pytest.mark.xfail(
_HAVE_FASTPARQUET and Version(fastparquet.__version__) > Version("2022.12"),
reason="fastparquet bug, see https://github.com/dask/fastparquet/issues/929",
)
@pytest.mark.skipif(using_copy_on_write(), reason="fastparquet writes into Index")
def test_timezone_aware_index(self, fp, timezone_aware_date_list):
def test_timezone_aware_index(self, fp, timezone_aware_date_list, request):
import fastparquet

if Version(fastparquet.__version__) < Version("2024.11.0"):
request.applymarker(
pytest.mark.xfail(
reason=(
"fastparquet bug, see "
"https://github.com/dask/fastparquet/issues/929"
),
)
)

idx = 5 * [timezone_aware_date_list]

df = pd.DataFrame(index=idx, data={"index_as_col": idx})
Expand Down

0 comments on commit c365dfc

Please sign in to comment.