Initial commit
This commit is contained in:
@@ -0,0 +1,16 @@
|
||||
# Licensed to the Apache Software Foundation (ASF) under one
|
||||
# or more contributor license agreements. See the NOTICE file
|
||||
# distributed with this work for additional information
|
||||
# regarding copyright ownership. The ASF licenses this file
|
||||
# to you under the Apache License, Version 2.0 (the
|
||||
# "License"); you may not use this file except in compliance
|
||||
# with the License. You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing,
|
||||
# software distributed under the License is distributed on an
|
||||
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||
# KIND, either express or implied. See the License for the
|
||||
# specific language governing permissions and limitations
|
||||
# under the License.
|
||||
Binary file not shown.
Binary file not shown.
Binary file not shown.
@@ -0,0 +1,529 @@
|
||||
# Licensed to the Apache Software Foundation (ASF) under one
|
||||
# or more contributor license agreements. See the NOTICE file
|
||||
# distributed with this work for additional information
|
||||
# regarding copyright ownership. The ASF licenses this file
|
||||
# to you under the Apache License, Version 2.0 (the
|
||||
# "License"); you may not use this file except in compliance
|
||||
# with the License. You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing,
|
||||
# software distributed under the License is distributed on an
|
||||
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||
# KIND, either express or implied. See the License for the
|
||||
# specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from datetime import datetime as dt
|
||||
import pyarrow as pa
|
||||
from pyarrow.vendored.version import Version
|
||||
import pytest
|
||||
|
||||
try:
|
||||
import numpy as np
|
||||
except ImportError:
|
||||
np = None
|
||||
|
||||
import pyarrow.interchange as pi
|
||||
from pyarrow.interchange.column import (
|
||||
_PyArrowColumn,
|
||||
ColumnNullType,
|
||||
DtypeKind,
|
||||
)
|
||||
from pyarrow.interchange.from_dataframe import _from_dataframe
|
||||
|
||||
try:
|
||||
import pandas as pd
|
||||
# import pandas.testing as tm
|
||||
except ImportError:
|
||||
pass
|
||||
|
||||
|
||||
@pytest.mark.parametrize("unit", ['s', 'ms', 'us', 'ns'])
|
||||
@pytest.mark.parametrize("tz", ['', 'America/New_York', '+07:30', '-04:30'])
|
||||
def test_datetime(unit, tz):
|
||||
dt_arr = [dt(2007, 7, 13), dt(2007, 7, 14), None]
|
||||
table = pa.table({"A": pa.array(dt_arr, type=pa.timestamp(unit, tz=tz))})
|
||||
col = table.__dataframe__().get_column_by_name("A")
|
||||
|
||||
assert col.size() == 3
|
||||
assert col.offset == 0
|
||||
assert col.null_count == 1
|
||||
assert col.dtype[0] == DtypeKind.DATETIME
|
||||
assert col.describe_null == (ColumnNullType.USE_BITMASK, 0)
|
||||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
["test_data", "kind"],
|
||||
[
|
||||
(["foo", "bar"], 21),
|
||||
([1.5, 2.5, 3.5], 2),
|
||||
([1, 2, 3, 4], 0),
|
||||
],
|
||||
)
|
||||
def test_array_to_pyarrowcolumn(test_data, kind):
|
||||
arr = pa.array(test_data)
|
||||
arr_column = _PyArrowColumn(arr)
|
||||
|
||||
assert arr_column._col == arr
|
||||
assert arr_column.size() == len(test_data)
|
||||
assert arr_column.dtype[0] == kind
|
||||
assert arr_column.num_chunks() == 1
|
||||
assert arr_column.null_count == 0
|
||||
assert arr_column.get_buffers()["validity"] is None
|
||||
assert len(list(arr_column.get_chunks())) == 1
|
||||
|
||||
for chunk in arr_column.get_chunks():
|
||||
assert chunk == arr_column
|
||||
|
||||
|
||||
def test_offset_of_sliced_array():
|
||||
arr = pa.array([1, 2, 3, 4])
|
||||
arr_sliced = arr.slice(2, 2)
|
||||
|
||||
table = pa.table([arr], names=["arr"])
|
||||
table_sliced = pa.table([arr_sliced], names=["arr_sliced"])
|
||||
|
||||
col = table_sliced.__dataframe__().get_column(0)
|
||||
assert col.offset == 2
|
||||
|
||||
result = _from_dataframe(table_sliced.__dataframe__())
|
||||
assert table_sliced.equals(result)
|
||||
assert not table.equals(result)
|
||||
|
||||
# pandas hardcodes offset to 0:
|
||||
# https://github.com/pandas-dev/pandas/blob/5c66e65d7b9fef47ccb585ce2fd0b3ea18dc82ea/pandas/core/interchange/from_dataframe.py#L247
|
||||
# so conversion to pandas can't be tested currently
|
||||
|
||||
# df = pandas_from_dataframe(table)
|
||||
# df_sliced = pandas_from_dataframe(table_sliced)
|
||||
|
||||
# tm.assert_series_equal(df["arr"][2:4], df_sliced["arr_sliced"],
|
||||
# check_index=False, check_names=False)
|
||||
|
||||
|
||||
@pytest.mark.pandas
|
||||
@pytest.mark.parametrize(
|
||||
"uint", [pa.uint8(), pa.uint16(), pa.uint32()]
|
||||
)
|
||||
@pytest.mark.parametrize(
|
||||
"int", [pa.int8(), pa.int16(), pa.int32(), pa.int64()]
|
||||
)
|
||||
@pytest.mark.parametrize(
|
||||
"float, np_float_str", [
|
||||
# (pa.float16(), np.float16), #not supported by pandas
|
||||
(pa.float32(), "float32"),
|
||||
(pa.float64(), "float64")
|
||||
]
|
||||
)
|
||||
def test_pandas_roundtrip(uint, int, float, np_float_str):
|
||||
if Version(pd.__version__) < Version("1.5.0"):
|
||||
pytest.skip("__dataframe__ added to pandas in 1.5.0")
|
||||
|
||||
arr = [1, 2, 3]
|
||||
table = pa.table(
|
||||
{
|
||||
"a": pa.array(arr, type=uint),
|
||||
"b": pa.array(arr, type=int),
|
||||
"c": pa.array(np.array(arr, dtype=np.dtype(np_float_str)), type=float),
|
||||
"d": [True, False, True],
|
||||
}
|
||||
)
|
||||
from pandas.api.interchange import (
|
||||
from_dataframe as pandas_from_dataframe
|
||||
)
|
||||
pandas_df = pandas_from_dataframe(table)
|
||||
result = pi.from_dataframe(pandas_df)
|
||||
assert table.equals(result)
|
||||
|
||||
table_protocol = table.__dataframe__()
|
||||
result_protocol = result.__dataframe__()
|
||||
|
||||
assert table_protocol.num_columns() == result_protocol.num_columns()
|
||||
assert table_protocol.num_rows() == result_protocol.num_rows()
|
||||
assert table_protocol.num_chunks() == result_protocol.num_chunks()
|
||||
assert table_protocol.column_names() == result_protocol.column_names()
|
||||
|
||||
|
||||
@pytest.mark.pandas
|
||||
def test_pandas_roundtrip_string():
|
||||
# See https://github.com/pandas-dev/pandas/issues/50554
|
||||
if Version(pd.__version__) < Version("1.6"):
|
||||
pytest.skip("Column.size() bug in pandas")
|
||||
|
||||
arr = ["a", "", "c"]
|
||||
table = pa.table({"a": pa.array(arr)})
|
||||
|
||||
from pandas.api.interchange import (
|
||||
from_dataframe as pandas_from_dataframe
|
||||
)
|
||||
|
||||
pandas_df = pandas_from_dataframe(table)
|
||||
result = pi.from_dataframe(pandas_df)
|
||||
|
||||
assert result["a"].to_pylist() == table["a"].to_pylist()
|
||||
assert pa.types.is_string(table["a"].type)
|
||||
assert pa.types.is_large_string(result["a"].type)
|
||||
|
||||
table_protocol = table.__dataframe__()
|
||||
result_protocol = result.__dataframe__()
|
||||
|
||||
assert table_protocol.num_columns() == result_protocol.num_columns()
|
||||
assert table_protocol.num_rows() == result_protocol.num_rows()
|
||||
assert table_protocol.num_chunks() == result_protocol.num_chunks()
|
||||
assert table_protocol.column_names() == result_protocol.column_names()
|
||||
|
||||
|
||||
@pytest.mark.pandas
|
||||
def test_pandas_roundtrip_large_string():
|
||||
# See https://github.com/pandas-dev/pandas/issues/50554
|
||||
if Version(pd.__version__) < Version("1.6"):
|
||||
pytest.skip("Column.size() bug in pandas")
|
||||
|
||||
arr = ["a", "", "c"]
|
||||
table = pa.table({"a_large": pa.array(arr, type=pa.large_string())})
|
||||
|
||||
from pandas.api.interchange import (
|
||||
from_dataframe as pandas_from_dataframe
|
||||
)
|
||||
|
||||
if Version(pd.__version__) >= Version("2.0.1"):
|
||||
pandas_df = pandas_from_dataframe(table)
|
||||
result = pi.from_dataframe(pandas_df)
|
||||
|
||||
assert result["a_large"].to_pylist() == table["a_large"].to_pylist()
|
||||
assert pa.types.is_large_string(table["a_large"].type)
|
||||
assert pa.types.is_large_string(result["a_large"].type)
|
||||
|
||||
table_protocol = table.__dataframe__()
|
||||
result_protocol = result.__dataframe__()
|
||||
|
||||
assert table_protocol.num_columns() == result_protocol.num_columns()
|
||||
assert table_protocol.num_rows() == result_protocol.num_rows()
|
||||
assert table_protocol.num_chunks() == result_protocol.num_chunks()
|
||||
assert table_protocol.column_names() == result_protocol.column_names()
|
||||
|
||||
else:
|
||||
# large string not supported by pandas implementation for
|
||||
# older versions of pandas
|
||||
# https://github.com/pandas-dev/pandas/issues/52795
|
||||
with pytest.raises(AssertionError):
|
||||
pandas_from_dataframe(table)
|
||||
|
||||
|
||||
@pytest.mark.pandas
|
||||
def test_pandas_roundtrip_string_with_missing():
|
||||
# See https://github.com/pandas-dev/pandas/issues/50554
|
||||
if Version(pd.__version__) < Version("1.6"):
|
||||
pytest.skip("Column.size() bug in pandas")
|
||||
|
||||
arr = ["a", "", "c", None]
|
||||
table = pa.table({"a": pa.array(arr),
|
||||
"a_large": pa.array(arr, type=pa.large_string())})
|
||||
|
||||
from pandas.api.interchange import (
|
||||
from_dataframe as pandas_from_dataframe
|
||||
)
|
||||
|
||||
if Version(pd.__version__) >= Version("2.0.2"):
|
||||
pandas_df = pandas_from_dataframe(table)
|
||||
result = pi.from_dataframe(pandas_df)
|
||||
|
||||
assert result["a"].to_pylist() == table["a"].to_pylist()
|
||||
assert pa.types.is_string(table["a"].type)
|
||||
assert pa.types.is_large_string(result["a"].type)
|
||||
|
||||
assert result["a_large"].to_pylist() == table["a_large"].to_pylist()
|
||||
assert pa.types.is_large_string(table["a_large"].type)
|
||||
assert pa.types.is_large_string(result["a_large"].type)
|
||||
else:
|
||||
# older versions of pandas do not have bitmask support
|
||||
# https://github.com/pandas-dev/pandas/issues/49888
|
||||
with pytest.raises(NotImplementedError):
|
||||
pandas_from_dataframe(table)
|
||||
|
||||
|
||||
@pytest.mark.pandas
|
||||
def test_pandas_roundtrip_categorical():
|
||||
if Version(pd.__version__) < Version("2.0.2"):
|
||||
pytest.skip("Bitmasks not supported in pandas interchange implementation")
|
||||
|
||||
arr = ["Mon", "Tue", "Mon", "Wed", "Mon", "Thu", "Fri", "Sat", None]
|
||||
table = pa.table(
|
||||
{"weekday": pa.array(arr).dictionary_encode()}
|
||||
)
|
||||
|
||||
from pandas.api.interchange import (
|
||||
from_dataframe as pandas_from_dataframe
|
||||
)
|
||||
pandas_df = pandas_from_dataframe(table)
|
||||
result = pi.from_dataframe(pandas_df)
|
||||
|
||||
assert result["weekday"].to_pylist() == table["weekday"].to_pylist()
|
||||
assert pa.types.is_dictionary(table["weekday"].type)
|
||||
assert pa.types.is_dictionary(result["weekday"].type)
|
||||
assert pa.types.is_string(table["weekday"].chunk(0).dictionary.type)
|
||||
assert pa.types.is_large_string(result["weekday"].chunk(0).dictionary.type)
|
||||
assert pa.types.is_int32(table["weekday"].chunk(0).indices.type)
|
||||
assert pa.types.is_int8(result["weekday"].chunk(0).indices.type)
|
||||
|
||||
table_protocol = table.__dataframe__()
|
||||
result_protocol = result.__dataframe__()
|
||||
|
||||
assert table_protocol.num_columns() == result_protocol.num_columns()
|
||||
assert table_protocol.num_rows() == result_protocol.num_rows()
|
||||
assert table_protocol.num_chunks() == result_protocol.num_chunks()
|
||||
assert table_protocol.column_names() == result_protocol.column_names()
|
||||
|
||||
col_table = table_protocol.get_column(0)
|
||||
col_result = result_protocol.get_column(0)
|
||||
|
||||
assert col_result.dtype[0] == DtypeKind.CATEGORICAL
|
||||
assert col_result.dtype[0] == col_table.dtype[0]
|
||||
assert col_result.size() == col_table.size()
|
||||
assert col_result.offset == col_table.offset
|
||||
|
||||
desc_cat_table = col_result.describe_categorical
|
||||
desc_cat_result = col_result.describe_categorical
|
||||
|
||||
assert desc_cat_table["is_ordered"] == desc_cat_result["is_ordered"]
|
||||
assert desc_cat_table["is_dictionary"] == desc_cat_result["is_dictionary"]
|
||||
assert isinstance(desc_cat_result["categories"]._col, pa.Array)
|
||||
|
||||
|
||||
@pytest.mark.pandas
|
||||
@pytest.mark.parametrize("unit", ['s', 'ms', 'us', 'ns'])
|
||||
def test_pandas_roundtrip_datetime(unit):
|
||||
if Version(pd.__version__) < Version("1.5.0"):
|
||||
pytest.skip("__dataframe__ added to pandas in 1.5.0")
|
||||
from datetime import datetime as dt
|
||||
|
||||
# timezones not included as they are not yet supported in
|
||||
# the pandas implementation
|
||||
dt_arr = [dt(2007, 7, 13), dt(2007, 7, 14), dt(2007, 7, 15)]
|
||||
table = pa.table({"a": pa.array(dt_arr, type=pa.timestamp(unit))})
|
||||
|
||||
if Version(pd.__version__) < Version("1.6"):
|
||||
# pandas < 2.0 always creates datetime64 in "ns"
|
||||
# resolution
|
||||
expected = pa.table({"a": pa.array(dt_arr, type=pa.timestamp('ns'))})
|
||||
else:
|
||||
expected = table
|
||||
|
||||
from pandas.api.interchange import (
|
||||
from_dataframe as pandas_from_dataframe
|
||||
)
|
||||
pandas_df = pandas_from_dataframe(table)
|
||||
result = pi.from_dataframe(pandas_df)
|
||||
|
||||
assert expected.equals(result)
|
||||
|
||||
expected_protocol = expected.__dataframe__()
|
||||
result_protocol = result.__dataframe__()
|
||||
|
||||
assert expected_protocol.num_columns() == result_protocol.num_columns()
|
||||
assert expected_protocol.num_rows() == result_protocol.num_rows()
|
||||
assert expected_protocol.num_chunks() == result_protocol.num_chunks()
|
||||
assert expected_protocol.column_names() == result_protocol.column_names()
|
||||
|
||||
|
||||
@pytest.mark.pandas
|
||||
@pytest.mark.parametrize(
|
||||
"np_float_str", ["float32", "float64"]
|
||||
)
|
||||
def test_pandas_to_pyarrow_with_missing(np_float_str):
|
||||
if Version(pd.__version__) < Version("1.5.0"):
|
||||
pytest.skip("__dataframe__ added to pandas in 1.5.0")
|
||||
|
||||
np_array = np.array([0, np.nan, 2], dtype=np.dtype(np_float_str))
|
||||
datetime_array = [None, dt(2007, 7, 14), dt(2007, 7, 15)]
|
||||
df = pd.DataFrame({
|
||||
# float, ColumnNullType.USE_NAN
|
||||
"a": np_array,
|
||||
# ColumnNullType.USE_SENTINEL
|
||||
"dt": np.array(datetime_array, dtype="datetime64[ns]")
|
||||
})
|
||||
expected = pa.table({
|
||||
"a": pa.array(np_array, from_pandas=True),
|
||||
"dt": pa.array(datetime_array, type=pa.timestamp("ns"))
|
||||
})
|
||||
result = pi.from_dataframe(df)
|
||||
|
||||
assert result.equals(expected)
|
||||
|
||||
|
||||
@pytest.mark.pandas
|
||||
def test_pandas_to_pyarrow_float16_with_missing():
|
||||
if Version(pd.__version__) < Version("1.5.0"):
|
||||
pytest.skip("__dataframe__ added to pandas in 1.5.0")
|
||||
|
||||
# np.float16 errors if ps.is_nan is used
|
||||
# pyarrow.lib.ArrowNotImplementedError: Function 'is_nan' has no kernel
|
||||
# matching input types (halffloat)
|
||||
np_array = np.array([0, np.nan, 2], dtype=np.float16)
|
||||
df = pd.DataFrame({"a": np_array})
|
||||
|
||||
with pytest.raises(NotImplementedError):
|
||||
pi.from_dataframe(df)
|
||||
|
||||
|
||||
@pytest.mark.numpy
|
||||
@pytest.mark.parametrize(
|
||||
"uint", [pa.uint8(), pa.uint16(), pa.uint32()]
|
||||
)
|
||||
@pytest.mark.parametrize(
|
||||
"int", [pa.int8(), pa.int16(), pa.int32(), pa.int64()]
|
||||
)
|
||||
@pytest.mark.parametrize(
|
||||
"float, np_float_str", [
|
||||
(pa.float16(), "float16"),
|
||||
(pa.float32(), "float32"),
|
||||
(pa.float64(), "float64")
|
||||
]
|
||||
)
|
||||
@pytest.mark.parametrize("unit", ['s', 'ms', 'us', 'ns'])
|
||||
@pytest.mark.parametrize("tz", ['America/New_York', '+07:30', '-04:30'])
|
||||
@pytest.mark.parametrize("offset, length", [(0, 3), (0, 2), (1, 2), (2, 1)])
|
||||
def test_pyarrow_roundtrip(uint, int, float, np_float_str,
|
||||
unit, tz, offset, length):
|
||||
|
||||
from datetime import datetime as dt
|
||||
arr = [1, 2, None]
|
||||
dt_arr = [dt(2007, 7, 13), None, dt(2007, 7, 15)]
|
||||
|
||||
table = pa.table(
|
||||
{
|
||||
"a": pa.array(arr, type=uint),
|
||||
"b": pa.array(arr, type=int),
|
||||
"c": pa.array(np.array(arr, dtype=np.dtype(np_float_str)),
|
||||
type=float, from_pandas=True),
|
||||
"d": [True, False, True],
|
||||
"e": [True, False, None],
|
||||
"f": ["a", None, "c"],
|
||||
"g": pa.array(dt_arr, type=pa.timestamp(unit, tz=tz))
|
||||
}
|
||||
)
|
||||
table = table.slice(offset, length)
|
||||
result = _from_dataframe(table.__dataframe__())
|
||||
|
||||
assert table.equals(result)
|
||||
|
||||
table_protocol = table.__dataframe__()
|
||||
result_protocol = result.__dataframe__()
|
||||
|
||||
assert table_protocol.num_columns() == result_protocol.num_columns()
|
||||
assert table_protocol.num_rows() == result_protocol.num_rows()
|
||||
assert table_protocol.num_chunks() == result_protocol.num_chunks()
|
||||
assert table_protocol.column_names() == result_protocol.column_names()
|
||||
|
||||
|
||||
@pytest.mark.parametrize("offset, length", [(0, 10), (0, 2), (7, 3), (2, 1)])
|
||||
def test_pyarrow_roundtrip_categorical(offset, length):
|
||||
arr = ["Mon", "Tue", "Mon", "Wed", "Mon", "Thu", "Fri", None, "Sun"]
|
||||
table = pa.table(
|
||||
{"weekday": pa.array(arr).dictionary_encode()}
|
||||
)
|
||||
table = table.slice(offset, length)
|
||||
result = _from_dataframe(table.__dataframe__())
|
||||
|
||||
assert table.equals(result)
|
||||
|
||||
table_protocol = table.__dataframe__()
|
||||
result_protocol = result.__dataframe__()
|
||||
|
||||
assert table_protocol.num_columns() == result_protocol.num_columns()
|
||||
assert table_protocol.num_rows() == result_protocol.num_rows()
|
||||
assert table_protocol.num_chunks() == result_protocol.num_chunks()
|
||||
assert table_protocol.column_names() == result_protocol.column_names()
|
||||
|
||||
col_table = table_protocol.get_column(0)
|
||||
col_result = result_protocol.get_column(0)
|
||||
|
||||
assert col_result.dtype[0] == DtypeKind.CATEGORICAL
|
||||
assert col_result.dtype[0] == col_table.dtype[0]
|
||||
assert col_result.size() == col_table.size()
|
||||
assert col_result.offset == col_table.offset
|
||||
|
||||
desc_cat_table = col_table.describe_categorical
|
||||
desc_cat_result = col_result.describe_categorical
|
||||
|
||||
assert desc_cat_table["is_ordered"] == desc_cat_result["is_ordered"]
|
||||
assert desc_cat_table["is_dictionary"] == desc_cat_result["is_dictionary"]
|
||||
assert isinstance(desc_cat_result["categories"]._col, pa.Array)
|
||||
|
||||
|
||||
@pytest.mark.large_memory
|
||||
def test_pyarrow_roundtrip_large_string():
|
||||
|
||||
data = np.array([b'x'*1024]*(3*1024**2), dtype='object') # 3GB bytes data
|
||||
arr = pa.array(data, type=pa.large_string())
|
||||
table = pa.table([arr], names=["large_string"])
|
||||
|
||||
result = _from_dataframe(table.__dataframe__())
|
||||
col = result.__dataframe__().get_column(0)
|
||||
|
||||
assert col.size() == 3*1024**2
|
||||
assert pa.types.is_large_string(table[0].type)
|
||||
assert pa.types.is_large_string(result[0].type)
|
||||
|
||||
assert table.equals(result)
|
||||
|
||||
|
||||
def test_nan_as_null():
|
||||
table = pa.table({"a": [1, 2, 3, 4]})
|
||||
with pytest.raises(RuntimeError):
|
||||
table.__dataframe__(nan_as_null=True)
|
||||
|
||||
|
||||
@pytest.mark.pandas
|
||||
def test_allow_copy_false():
|
||||
if Version(pd.__version__) < Version("1.5.0"):
|
||||
pytest.skip("__dataframe__ added to pandas in 1.5.0")
|
||||
|
||||
# Test that an error is raised when a copy is needed
|
||||
# to create a bitmask
|
||||
|
||||
df = pd.DataFrame({"a": [0, 1.0, 2.0]})
|
||||
with pytest.raises(RuntimeError):
|
||||
pi.from_dataframe(df, allow_copy=False)
|
||||
|
||||
df = pd.DataFrame({
|
||||
"dt": [None, dt(2007, 7, 14), dt(2007, 7, 15)]
|
||||
})
|
||||
with pytest.raises(RuntimeError):
|
||||
pi.from_dataframe(df, allow_copy=False)
|
||||
|
||||
|
||||
@pytest.mark.pandas
|
||||
def test_allow_copy_false_bool_categorical():
|
||||
if Version(pd.__version__) < Version("1.5.0"):
|
||||
pytest.skip("__dataframe__ added to pandas in 1.5.0")
|
||||
|
||||
# Test that an error is raised for boolean
|
||||
# and categorical dtype (copy is always made)
|
||||
|
||||
df = pd.DataFrame({"a": [None, False, True]})
|
||||
with pytest.raises(RuntimeError):
|
||||
pi.from_dataframe(df, allow_copy=False)
|
||||
|
||||
df = pd.DataFrame({"a": [True, False, True]})
|
||||
with pytest.raises(RuntimeError):
|
||||
pi.from_dataframe(df, allow_copy=False)
|
||||
|
||||
df = pd.DataFrame({"weekday": ["a", "b", None]})
|
||||
df = df.astype("category")
|
||||
with pytest.raises(RuntimeError):
|
||||
pi.from_dataframe(df, allow_copy=False)
|
||||
|
||||
df = pd.DataFrame({"weekday": ["a", "b", "c"]})
|
||||
df = df.astype("category")
|
||||
with pytest.raises(RuntimeError):
|
||||
pi.from_dataframe(df, allow_copy=False)
|
||||
|
||||
|
||||
def test_empty_dataframe():
|
||||
schema = pa.schema([('col1', pa.int8())])
|
||||
df = pa.table([[]], schema=schema)
|
||||
dfi = df.__dataframe__()
|
||||
assert pi.from_dataframe(dfi) == df
|
||||
@@ -0,0 +1,294 @@
|
||||
# Licensed to the Apache Software Foundation (ASF) under one
|
||||
# or more contributor license agreements. See the NOTICE file
|
||||
# distributed with this work for additional information
|
||||
# regarding copyright ownership. The ASF licenses this file
|
||||
# to you under the Apache License, Version 2.0 (the
|
||||
# "License"); you may not use this file except in compliance
|
||||
# with the License. You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing,
|
||||
# software distributed under the License is distributed on an
|
||||
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||
# KIND, either express or implied. See the License for the
|
||||
# specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import ctypes
|
||||
import hypothesis as h
|
||||
import hypothesis.strategies as st
|
||||
|
||||
import pytest
|
||||
try:
|
||||
import numpy as np
|
||||
except ImportError:
|
||||
np = None
|
||||
import pyarrow as pa
|
||||
import pyarrow.tests.strategies as past
|
||||
|
||||
|
||||
all_types = st.deferred(
|
||||
lambda: (
|
||||
past.signed_integer_types |
|
||||
past.unsigned_integer_types |
|
||||
past.floating_types |
|
||||
past.bool_type |
|
||||
past.string_type |
|
||||
past.large_string_type
|
||||
)
|
||||
)
|
||||
|
||||
|
||||
# datetime is tested in test_extra.py
|
||||
# dictionary is tested in test_categorical()
|
||||
@pytest.mark.numpy
|
||||
@h.settings(suppress_health_check=(h.HealthCheck.too_slow,))
|
||||
@h.given(past.arrays(all_types, size=3))
|
||||
def test_dtypes(arr):
|
||||
table = pa.table([arr], names=["a"])
|
||||
df = table.__dataframe__()
|
||||
|
||||
null_count = df.get_column(0).null_count
|
||||
assert null_count == arr.null_count
|
||||
assert isinstance(null_count, int)
|
||||
assert df.get_column(0).size() == 3
|
||||
assert df.get_column(0).offset == 0
|
||||
|
||||
|
||||
@pytest.mark.numpy
|
||||
@pytest.mark.parametrize(
|
||||
"uint, uint_bw",
|
||||
[
|
||||
(pa.uint8(), 8),
|
||||
(pa.uint16(), 16),
|
||||
(pa.uint32(), 32)
|
||||
]
|
||||
)
|
||||
@pytest.mark.parametrize(
|
||||
"int, int_bw", [
|
||||
(pa.int8(), 8),
|
||||
(pa.int16(), 16),
|
||||
(pa.int32(), 32),
|
||||
(pa.int64(), 64)
|
||||
]
|
||||
)
|
||||
@pytest.mark.parametrize(
|
||||
"float, float_bw, np_float_str", [
|
||||
(pa.float16(), 16, "float16"),
|
||||
(pa.float32(), 32, "float32"),
|
||||
(pa.float64(), 64, "float64")
|
||||
]
|
||||
)
|
||||
@pytest.mark.parametrize("unit", ['s', 'ms', 'us', 'ns'])
|
||||
@pytest.mark.parametrize("tz", ['', 'America/New_York', '+07:30', '-04:30'])
|
||||
@pytest.mark.parametrize("use_batch", [False, True])
|
||||
def test_mixed_dtypes(uint, uint_bw, int, int_bw,
|
||||
float, float_bw, np_float_str, unit, tz,
|
||||
use_batch):
|
||||
from datetime import datetime as dt
|
||||
arr = [1, 2, 3]
|
||||
dt_arr = [dt(2007, 7, 13), dt(2007, 7, 14), dt(2007, 7, 15)]
|
||||
table = pa.table(
|
||||
{
|
||||
"a": pa.array(arr, type=uint),
|
||||
"b": pa.array(arr, type=int),
|
||||
"c": pa.array(np.array(arr, dtype=np.dtype(np_float_str)), type=float),
|
||||
"d": [True, False, True],
|
||||
"e": ["a", "", "c"],
|
||||
"f": pa.array(dt_arr, type=pa.timestamp(unit, tz=tz))
|
||||
}
|
||||
)
|
||||
if use_batch:
|
||||
table = table.to_batches()[0]
|
||||
df = table.__dataframe__()
|
||||
# 0 = DtypeKind.INT, 1 = DtypeKind.UINT, 2 = DtypeKind.FLOAT,
|
||||
# 20 = DtypeKind.BOOL, 21 = DtypeKind.STRING, 22 = DtypeKind.DATETIME
|
||||
# see DtypeKind class in column.py
|
||||
columns = {"a": 1, "b": 0, "c": 2, "d": 20, "e": 21, "f": 22}
|
||||
|
||||
for column, kind in columns.items():
|
||||
col = df.get_column_by_name(column)
|
||||
|
||||
assert col.null_count == 0
|
||||
assert col.size() == 3
|
||||
assert col.offset == 0
|
||||
assert col.dtype[0] == kind
|
||||
|
||||
assert df.get_column_by_name("a").dtype[1] == uint_bw
|
||||
assert df.get_column_by_name("b").dtype[1] == int_bw
|
||||
assert df.get_column_by_name("c").dtype[1] == float_bw
|
||||
|
||||
|
||||
def test_na_float():
|
||||
table = pa.table({"a": [1.0, None, 2.0]})
|
||||
df = table.__dataframe__()
|
||||
col = df.get_column_by_name("a")
|
||||
assert col.null_count == 1
|
||||
assert isinstance(col.null_count, int)
|
||||
|
||||
|
||||
def test_noncategorical():
|
||||
table = pa.table({"a": [1, 2, 3]})
|
||||
df = table.__dataframe__()
|
||||
col = df.get_column_by_name("a")
|
||||
with pytest.raises(TypeError, match=".*categorical.*"):
|
||||
col.describe_categorical
|
||||
|
||||
|
||||
@pytest.mark.parametrize("use_batch", [False, True])
|
||||
def test_categorical(use_batch):
|
||||
import pyarrow as pa
|
||||
arr = ["Mon", "Tue", "Mon", "Wed", "Mon", "Thu", "Fri", "Sat", None]
|
||||
table = pa.table(
|
||||
{"weekday": pa.array(arr).dictionary_encode()}
|
||||
)
|
||||
if use_batch:
|
||||
table = table.to_batches()[0]
|
||||
|
||||
col = table.__dataframe__().get_column_by_name("weekday")
|
||||
categorical = col.describe_categorical
|
||||
assert isinstance(categorical["is_ordered"], bool)
|
||||
assert isinstance(categorical["is_dictionary"], bool)
|
||||
|
||||
|
||||
@pytest.mark.parametrize("use_batch", [False, True])
|
||||
def test_dataframe(use_batch):
|
||||
n = pa.chunked_array([[2, 2, 4], [4, 5, 100]])
|
||||
a = pa.chunked_array([["Flamingo", "Parrot", "Cow"],
|
||||
["Horse", "Brittle stars", "Centipede"]])
|
||||
table = pa.table([n, a], names=['n_legs', 'animals'])
|
||||
if use_batch:
|
||||
table = table.combine_chunks().to_batches()[0]
|
||||
df = table.__dataframe__()
|
||||
|
||||
assert df.num_columns() == 2
|
||||
assert df.num_rows() == 6
|
||||
if use_batch:
|
||||
assert df.num_chunks() == 1
|
||||
else:
|
||||
assert df.num_chunks() == 2
|
||||
assert list(df.column_names()) == ['n_legs', 'animals']
|
||||
assert list(df.select_columns((1,)).column_names()) == list(
|
||||
df.select_columns_by_name(("animals",)).column_names()
|
||||
)
|
||||
|
||||
|
||||
@pytest.mark.parametrize("use_batch", [False, True])
|
||||
@pytest.mark.parametrize(["size", "n_chunks"], [(10, 3), (12, 3), (12, 5)])
|
||||
def test_df_get_chunks(use_batch, size, n_chunks):
|
||||
table = pa.table({"x": list(range(size))})
|
||||
if use_batch:
|
||||
table = table.to_batches()[0]
|
||||
df = table.__dataframe__()
|
||||
chunks = list(df.get_chunks(n_chunks))
|
||||
assert len(chunks) == n_chunks
|
||||
assert sum(chunk.num_rows() for chunk in chunks) == size
|
||||
|
||||
|
||||
@pytest.mark.parametrize("use_batch", [False, True])
|
||||
@pytest.mark.parametrize(["size", "n_chunks"], [(10, 3), (12, 3), (12, 5)])
|
||||
def test_column_get_chunks(use_batch, size, n_chunks):
|
||||
table = pa.table({"x": list(range(size))})
|
||||
if use_batch:
|
||||
table = table.to_batches()[0]
|
||||
df = table.__dataframe__()
|
||||
chunks = list(df.get_column(0).get_chunks(n_chunks))
|
||||
assert len(chunks) == n_chunks
|
||||
assert sum(chunk.size() for chunk in chunks) == size
|
||||
|
||||
|
||||
@pytest.mark.pandas
|
||||
@pytest.mark.parametrize(
|
||||
"uint", [pa.uint8(), pa.uint16(), pa.uint32()]
|
||||
)
|
||||
@pytest.mark.parametrize(
|
||||
"int", [pa.int8(), pa.int16(), pa.int32(), pa.int64()]
|
||||
)
|
||||
@pytest.mark.parametrize(
|
||||
"float, np_float_str", [
|
||||
(pa.float16(), "float16"),
|
||||
(pa.float32(), "float32"),
|
||||
(pa.float64(), "float64")
|
||||
]
|
||||
)
|
||||
@pytest.mark.parametrize("use_batch", [False, True])
|
||||
def test_get_columns(uint, int, float, np_float_str, use_batch):
|
||||
arr = [[1, 2, 3], [4, 5]]
|
||||
arr_float = np.array([1, 2, 3, 4, 5], dtype=np.dtype(np_float_str))
|
||||
table = pa.table(
|
||||
{
|
||||
"a": pa.chunked_array(arr, type=uint),
|
||||
"b": pa.chunked_array(arr, type=int),
|
||||
"c": pa.array(arr_float, type=float)
|
||||
}
|
||||
)
|
||||
if use_batch:
|
||||
table = table.combine_chunks().to_batches()[0]
|
||||
df = table.__dataframe__()
|
||||
for col in df.get_columns():
|
||||
assert col.size() == 5
|
||||
assert col.num_chunks() == 1
|
||||
|
||||
# 0 = DtypeKind.INT, 1 = DtypeKind.UINT, 2 = DtypeKind.FLOAT,
|
||||
# see DtypeKind class in column.py
|
||||
assert df.get_column(0).dtype[0] == 1 # UINT
|
||||
assert df.get_column(1).dtype[0] == 0 # INT
|
||||
assert df.get_column(2).dtype[0] == 2 # FLOAT
|
||||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
"int", [pa.int8(), pa.int16(), pa.int32(), pa.int64()]
|
||||
)
|
||||
@pytest.mark.parametrize("use_batch", [False, True])
|
||||
def test_buffer(int, use_batch):
|
||||
arr = [0, 1, -1]
|
||||
table = pa.table({"a": pa.array(arr, type=int)})
|
||||
if use_batch:
|
||||
table = table.to_batches()[0]
|
||||
df = table.__dataframe__()
|
||||
col = df.get_column(0)
|
||||
buf = col.get_buffers()
|
||||
|
||||
dataBuf, dataDtype = buf["data"]
|
||||
|
||||
assert dataBuf.bufsize > 0
|
||||
assert dataBuf.ptr != 0
|
||||
device, _ = dataBuf.__dlpack_device__()
|
||||
|
||||
# 0 = DtypeKind.INT
|
||||
# see DtypeKind class in column.py
|
||||
assert dataDtype[0] == 0
|
||||
|
||||
if device == 1: # CPU-only as we're going to directly read memory here
|
||||
bitwidth = dataDtype[1]
|
||||
ctype = {
|
||||
8: ctypes.c_int8,
|
||||
16: ctypes.c_int16,
|
||||
32: ctypes.c_int32,
|
||||
64: ctypes.c_int64,
|
||||
}[bitwidth]
|
||||
|
||||
for idx, truth in enumerate(arr):
|
||||
val = ctype.from_address(dataBuf.ptr + idx * (bitwidth // 8)).value
|
||||
assert val == truth, f"Buffer at index {idx} mismatch"
|
||||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
"indices_type, bitwidth, f_string", [
|
||||
(pa.int8(), 8, "c"),
|
||||
(pa.int16(), 16, "s"),
|
||||
(pa.int32(), 32, "i"),
|
||||
(pa.int64(), 64, "l")
|
||||
]
|
||||
)
|
||||
def test_categorical_dtype(indices_type, bitwidth, f_string):
|
||||
type = pa.dictionary(indices_type, pa.string())
|
||||
arr = pa.array(["a", "b", None, "d"], type)
|
||||
table = pa.table({'a': arr})
|
||||
|
||||
df = table.__dataframe__()
|
||||
col = df.get_column(0)
|
||||
assert col.dtype[0] == 23 # <DtypeKind.CATEGORICAL: 23>
|
||||
assert col.dtype[1] == bitwidth
|
||||
assert col.dtype[2] == f_string
|
||||
Reference in New Issue
Block a user