|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import math |
|
|
import time |
|
|
import uuid |
|
|
from urllib.parse import urlparse |
|
|
|
|
|
import pyarrow.parquet as pq |
|
|
import pytest |
|
|
from hive_metastore.ttypes import LockRequest, LockResponse, LockState, UnlockRequest |
|
|
from pyarrow.fs import S3FileSystem |
|
|
from pydantic_core import ValidationError |
|
|
|
|
|
from pyiceberg.catalog import Catalog |
|
|
from pyiceberg.catalog.hive import HiveCatalog, _HiveClient |
|
|
from pyiceberg.exceptions import CommitFailedException, NoSuchTableError |
|
|
from pyiceberg.expressions import ( |
|
|
And, |
|
|
EqualTo, |
|
|
GreaterThanOrEqual, |
|
|
IsNaN, |
|
|
LessThan, |
|
|
NotEqualTo, |
|
|
NotNaN, |
|
|
) |
|
|
from pyiceberg.io.pyarrow import pyarrow_to_schema |
|
|
from pyiceberg.schema import Schema |
|
|
from pyiceberg.table import Table |
|
|
from pyiceberg.types import ( |
|
|
BooleanType, |
|
|
IntegerType, |
|
|
NestedField, |
|
|
StringType, |
|
|
TimestampType, |
|
|
) |
|
|
from pyiceberg.utils.concurrent import ExecutorFactory |
|
|
|
|
|
DEFAULT_PROPERTIES = {"write.parquet.compression-codec": "zstd"} |
|
|
|
|
|
|
|
|
TABLE_NAME = ("default", "t1") |
|
|
|
|
|
|
|
|
def create_table(catalog: Catalog) -> Table: |
|
|
try: |
|
|
catalog.drop_table(TABLE_NAME) |
|
|
except NoSuchTableError: |
|
|
pass |
|
|
|
|
|
schema = Schema( |
|
|
NestedField(field_id=1, name="str", field_type=StringType(), required=False), |
|
|
NestedField(field_id=2, name="int", field_type=IntegerType(), required=True), |
|
|
NestedField(field_id=3, name="bool", field_type=BooleanType(), required=False), |
|
|
NestedField(field_id=4, name="datetime", field_type=TimestampType(), required=False), |
|
|
) |
|
|
|
|
|
return catalog.create_table(identifier=TABLE_NAME, schema=schema) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_table_properties(catalog: Catalog) -> None: |
|
|
table = create_table(catalog) |
|
|
|
|
|
assert table.properties == DEFAULT_PROPERTIES |
|
|
|
|
|
with table.transaction() as transaction: |
|
|
transaction.set_properties(abc="🤪") |
|
|
assert table.properties == dict(abc="🤪", **DEFAULT_PROPERTIES) |
|
|
|
|
|
with table.transaction() as transaction: |
|
|
transaction.remove_properties("abc") |
|
|
assert table.properties == DEFAULT_PROPERTIES |
|
|
|
|
|
table = table.transaction().set_properties(abc="def").commit_transaction() |
|
|
assert table.properties == dict(abc="def", **DEFAULT_PROPERTIES) |
|
|
|
|
|
table = table.transaction().remove_properties("abc").commit_transaction() |
|
|
assert table.properties == DEFAULT_PROPERTIES |
|
|
|
|
|
table = table.transaction().set_properties(abc=123).commit_transaction() |
|
|
|
|
|
assert table.properties == dict(abc="123", **DEFAULT_PROPERTIES) |
|
|
|
|
|
with pytest.raises(ValidationError) as exc_info: |
|
|
table.transaction().set_properties(property_name=None).commit_transaction() |
|
|
assert "None type is not a supported value in properties: property_name" in str(exc_info.value) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_table_properties_dict(catalog: Catalog) -> None: |
|
|
table = create_table(catalog) |
|
|
|
|
|
assert table.properties == DEFAULT_PROPERTIES |
|
|
|
|
|
with table.transaction() as transaction: |
|
|
transaction.set_properties({"abc": "🤪"}) |
|
|
assert table.properties == dict({"abc": "🤪"}, **DEFAULT_PROPERTIES) |
|
|
|
|
|
with table.transaction() as transaction: |
|
|
transaction.remove_properties("abc") |
|
|
assert table.properties == DEFAULT_PROPERTIES |
|
|
|
|
|
table = table.transaction().set_properties({"abc": "def"}).commit_transaction() |
|
|
assert table.properties == dict({"abc": "def"}, **DEFAULT_PROPERTIES) |
|
|
|
|
|
table = table.transaction().remove_properties("abc").commit_transaction() |
|
|
assert table.properties == DEFAULT_PROPERTIES |
|
|
|
|
|
table = table.transaction().set_properties({"abc": 123}).commit_transaction() |
|
|
|
|
|
assert table.properties == dict({"abc": "123"}, **DEFAULT_PROPERTIES) |
|
|
|
|
|
with pytest.raises(ValidationError) as exc_info: |
|
|
table.transaction().set_properties({"property_name": None}).commit_transaction() |
|
|
assert "None type is not a supported value in properties: property_name" in str(exc_info.value) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_table_properties_error(catalog: Catalog) -> None: |
|
|
table = create_table(catalog) |
|
|
properties = {"abc": "def"} |
|
|
with pytest.raises(ValueError) as e: |
|
|
table.transaction().set_properties(properties, abc="def").commit_transaction() |
|
|
assert "Cannot pass both properties and kwargs" in str(e.value) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_pyarrow_nan(catalog: Catalog) -> None: |
|
|
table_test_null_nan = catalog.load_table("default.test_null_nan") |
|
|
arrow_table = table_test_null_nan.scan(row_filter=IsNaN("col_numeric"), selected_fields=("idx", "col_numeric")).to_arrow() |
|
|
assert len(arrow_table) == 1 |
|
|
assert arrow_table["idx"][0].as_py() == 1 |
|
|
assert math.isnan(arrow_table["col_numeric"][0].as_py()) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_pyarrow_nan_rewritten(catalog: Catalog) -> None: |
|
|
table_test_null_nan_rewritten = catalog.load_table("default.test_null_nan_rewritten") |
|
|
arrow_table = table_test_null_nan_rewritten.scan( |
|
|
row_filter=IsNaN("col_numeric"), selected_fields=("idx", "col_numeric") |
|
|
).to_arrow() |
|
|
assert len(arrow_table) == 1 |
|
|
assert arrow_table["idx"][0].as_py() == 1 |
|
|
assert math.isnan(arrow_table["col_numeric"][0].as_py()) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
@pytest.mark.skip(reason="Fixing issues with NaN's: https://github.com/apache/arrow/issues/34162") |
|
|
def test_pyarrow_not_nan_count(catalog: Catalog) -> None: |
|
|
table_test_null_nan = catalog.load_table("default.test_null_nan") |
|
|
not_nan = table_test_null_nan.scan(row_filter=NotNaN("col_numeric"), selected_fields=("idx",)).to_arrow() |
|
|
assert len(not_nan) == 2 |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_duckdb_nan(catalog: Catalog) -> None: |
|
|
table_test_null_nan_rewritten = catalog.load_table("default.test_null_nan_rewritten") |
|
|
con = table_test_null_nan_rewritten.scan().to_duckdb("table_test_null_nan") |
|
|
result = con.query("SELECT idx, col_numeric FROM table_test_null_nan WHERE isnan(col_numeric)").fetchone() |
|
|
assert result[0] == 1 |
|
|
assert math.isnan(result[1]) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_pyarrow_limit(catalog: Catalog) -> None: |
|
|
table_test_limit = catalog.load_table("default.test_limit") |
|
|
limited_result = table_test_limit.scan(selected_fields=("idx",), limit=1).to_arrow() |
|
|
assert len(limited_result) == 1 |
|
|
|
|
|
empty_result = table_test_limit.scan(selected_fields=("idx",), limit=0).to_arrow() |
|
|
assert len(empty_result) == 0 |
|
|
|
|
|
full_result = table_test_limit.scan(selected_fields=("idx",), limit=999).to_arrow() |
|
|
assert len(full_result) == 10 |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.filterwarnings("ignore") |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_daft_nan(catalog: Catalog) -> None: |
|
|
table_test_null_nan_rewritten = catalog.load_table("default.test_null_nan_rewritten") |
|
|
df = table_test_null_nan_rewritten.to_daft() |
|
|
assert df.count_rows() == 3 |
|
|
assert math.isnan(df.to_pydict()["col_numeric"][0]) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_daft_nan_rewritten(catalog: Catalog) -> None: |
|
|
table_test_null_nan_rewritten = catalog.load_table("default.test_null_nan_rewritten") |
|
|
df = table_test_null_nan_rewritten.to_daft() |
|
|
df = df.where(df["col_numeric"].float.is_nan()) |
|
|
df = df.select("idx", "col_numeric") |
|
|
assert df.count_rows() == 1 |
|
|
assert df.to_pydict()["idx"][0] == 1 |
|
|
assert math.isnan(df.to_pydict()["col_numeric"][0]) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.filterwarnings("ignore") |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_ray_nan(catalog: Catalog) -> None: |
|
|
table_test_null_nan_rewritten = catalog.load_table("default.test_null_nan_rewritten") |
|
|
ray_dataset = table_test_null_nan_rewritten.scan().to_ray() |
|
|
assert ray_dataset.count() == 3 |
|
|
assert math.isnan(ray_dataset.take()[0]["col_numeric"]) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_ray_nan_rewritten(catalog: Catalog) -> None: |
|
|
table_test_null_nan_rewritten = catalog.load_table("default.test_null_nan_rewritten") |
|
|
ray_dataset = table_test_null_nan_rewritten.scan( |
|
|
row_filter=IsNaN("col_numeric"), selected_fields=("idx", "col_numeric") |
|
|
).to_ray() |
|
|
assert ray_dataset.count() == 1 |
|
|
assert ray_dataset.take()[0]["idx"] == 1 |
|
|
assert math.isnan(ray_dataset.take()[0]["col_numeric"]) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
@pytest.mark.skip(reason="Fixing issues with NaN's: https://github.com/apache/arrow/issues/34162") |
|
|
def test_ray_not_nan_count(catalog: Catalog) -> None: |
|
|
table_test_null_nan_rewritten = catalog.load_table("default.test_null_nan_rewritten") |
|
|
ray_dataset = table_test_null_nan_rewritten.scan(row_filter=NotNaN("col_numeric"), selected_fields=("idx",)).to_ray() |
|
|
assert ray_dataset.count() == 2 |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_ray_all_types(catalog: Catalog) -> None: |
|
|
table_test_all_types = catalog.load_table("default.test_all_types") |
|
|
ray_dataset = table_test_all_types.scan().to_ray() |
|
|
pandas_dataframe = table_test_all_types.scan().to_pandas() |
|
|
assert ray_dataset.count() == pandas_dataframe.shape[0] |
|
|
assert pandas_dataframe.equals(ray_dataset.to_pandas()) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_pyarrow_to_iceberg_all_types(catalog: Catalog) -> None: |
|
|
table_test_all_types = catalog.load_table("default.test_all_types") |
|
|
fs = S3FileSystem( |
|
|
endpoint_override=catalog.properties["s3.endpoint"], |
|
|
access_key=catalog.properties["s3.access-key-id"], |
|
|
secret_key=catalog.properties["s3.secret-access-key"], |
|
|
) |
|
|
data_file_paths = [task.file.file_path for task in table_test_all_types.scan().plan_files()] |
|
|
for data_file_path in data_file_paths: |
|
|
uri = urlparse(data_file_path) |
|
|
with fs.open_input_file(f"{uri.netloc}{uri.path}") as fout: |
|
|
parquet_schema = pq.read_schema(fout) |
|
|
stored_iceberg_schema = Schema.model_validate_json(parquet_schema.metadata.get(b"iceberg.schema")) |
|
|
converted_iceberg_schema = pyarrow_to_schema(parquet_schema) |
|
|
assert converted_iceberg_schema == stored_iceberg_schema |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_pyarrow_deletes(catalog: Catalog) -> None: |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
test_positional_mor_deletes = catalog.load_table("default.test_positional_mor_deletes") |
|
|
arrow_table = test_positional_mor_deletes.scan().to_arrow() |
|
|
assert arrow_table["number"].to_pylist() == [1, 2, 3, 4, 5, 6, 7, 8, 10, 11, 12] |
|
|
|
|
|
|
|
|
arrow_table = test_positional_mor_deletes.scan( |
|
|
row_filter=And(GreaterThanOrEqual("letter", "e"), LessThan("letter", "k")) |
|
|
).to_arrow() |
|
|
assert arrow_table["number"].to_pylist() == [5, 6, 7, 8, 10] |
|
|
|
|
|
|
|
|
arrow_table = test_positional_mor_deletes.scan( |
|
|
row_filter=And(GreaterThanOrEqual("letter", "e"), LessThan("letter", "k")), limit=1 |
|
|
).to_arrow() |
|
|
assert arrow_table["number"].to_pylist() == [5] |
|
|
|
|
|
|
|
|
arrow_table = test_positional_mor_deletes.scan(limit=3).to_arrow() |
|
|
assert arrow_table["number"].to_pylist() == [1, 2, 3] |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_pyarrow_deletes_double(catalog: Catalog) -> None: |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
test_positional_mor_double_deletes = catalog.load_table("default.test_positional_mor_double_deletes") |
|
|
arrow_table = test_positional_mor_double_deletes.scan().to_arrow() |
|
|
assert arrow_table["number"].to_pylist() == [1, 2, 3, 4, 5, 7, 8, 10, 11, 12] |
|
|
|
|
|
|
|
|
arrow_table = test_positional_mor_double_deletes.scan( |
|
|
row_filter=And(GreaterThanOrEqual("letter", "e"), LessThan("letter", "k")) |
|
|
).to_arrow() |
|
|
assert arrow_table["number"].to_pylist() == [5, 7, 8, 10] |
|
|
|
|
|
|
|
|
arrow_table = test_positional_mor_double_deletes.scan( |
|
|
row_filter=And(GreaterThanOrEqual("letter", "e"), LessThan("letter", "k")), limit=1 |
|
|
).to_arrow() |
|
|
assert arrow_table["number"].to_pylist() == [5] |
|
|
|
|
|
|
|
|
arrow_table = test_positional_mor_double_deletes.scan(limit=8).to_arrow() |
|
|
assert arrow_table["number"].to_pylist() == [1, 2, 3, 4, 5, 7, 8, 10] |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_partitioned_tables(catalog: Catalog) -> None: |
|
|
for table_name, predicate in [ |
|
|
("test_partitioned_by_identity", "ts >= '2023-03-05T00:00:00+00:00'"), |
|
|
("test_partitioned_by_years", "dt >= '2023-03-05'"), |
|
|
("test_partitioned_by_months", "dt >= '2023-03-05'"), |
|
|
("test_partitioned_by_days", "ts >= '2023-03-05T00:00:00+00:00'"), |
|
|
("test_partitioned_by_hours", "ts >= '2023-03-05T00:00:00+00:00'"), |
|
|
("test_partitioned_by_truncate", "letter >= 'e'"), |
|
|
("test_partitioned_by_bucket", "number >= '5'"), |
|
|
]: |
|
|
table = catalog.load_table(f"default.{table_name}") |
|
|
arrow_table = table.scan(selected_fields=("number",), row_filter=predicate).to_arrow() |
|
|
assert set(arrow_table["number"].to_pylist()) == {5, 6, 7, 8, 9, 10, 11, 12}, f"Table {table_name}, predicate {predicate}" |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_unpartitioned_uuid_table(catalog: Catalog) -> None: |
|
|
unpartitioned_uuid = catalog.load_table("default.test_uuid_and_fixed_unpartitioned") |
|
|
arrow_table_eq = unpartitioned_uuid.scan(row_filter="uuid_col == '102cb62f-e6f8-4eb0-9973-d9b012ff0967'").to_arrow() |
|
|
assert arrow_table_eq["uuid_col"].to_pylist() == [uuid.UUID("102cb62f-e6f8-4eb0-9973-d9b012ff0967").bytes] |
|
|
|
|
|
arrow_table_neq = unpartitioned_uuid.scan( |
|
|
row_filter="uuid_col != '102cb62f-e6f8-4eb0-9973-d9b012ff0967' and uuid_col != '639cccce-c9d2-494a-a78c-278ab234f024'" |
|
|
).to_arrow() |
|
|
assert arrow_table_neq["uuid_col"].to_pylist() == [ |
|
|
uuid.UUID("ec33e4b2-a834-4cc3-8c4a-a1d3bfc2f226").bytes, |
|
|
uuid.UUID("c1b0d8e0-0b0e-4b1e-9b0a-0e0b0d0c0a0b").bytes, |
|
|
uuid.UUID("923dae77-83d6-47cd-b4b0-d383e64ee57e").bytes, |
|
|
] |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_unpartitioned_fixed_table(catalog: Catalog) -> None: |
|
|
fixed_table = catalog.load_table("default.test_uuid_and_fixed_unpartitioned") |
|
|
arrow_table_eq = fixed_table.scan(row_filter=EqualTo("fixed_col", b"1234567890123456789012345")).to_arrow() |
|
|
assert arrow_table_eq["fixed_col"].to_pylist() == [b"1234567890123456789012345"] |
|
|
|
|
|
arrow_table_neq = fixed_table.scan( |
|
|
row_filter=And( |
|
|
NotEqualTo("fixed_col", b"1234567890123456789012345"), NotEqualTo("uuid_col", "c1b0d8e0-0b0e-4b1e-9b0a-0e0b0d0c0a0b") |
|
|
) |
|
|
).to_arrow() |
|
|
assert arrow_table_neq["fixed_col"].to_pylist() == [ |
|
|
b"1231231231231231231231231", |
|
|
b"12345678901234567ass12345", |
|
|
b"qweeqwwqq1231231231231111", |
|
|
] |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_scan_tag(catalog: Catalog) -> None: |
|
|
test_positional_mor_deletes = catalog.load_table("default.test_positional_mor_deletes") |
|
|
arrow_table = test_positional_mor_deletes.scan().use_ref("tag_12").to_arrow() |
|
|
assert arrow_table["number"].to_pylist() == [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12] |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_scan_branch(catalog: Catalog) -> None: |
|
|
test_positional_mor_deletes = catalog.load_table("default.test_positional_mor_deletes") |
|
|
arrow_table = test_positional_mor_deletes.scan().use_ref("without_5").to_arrow() |
|
|
assert arrow_table["number"].to_pylist() == [1, 2, 3, 4, 6, 7, 8, 9, 10, 11, 12] |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_filter_on_new_column(catalog: Catalog) -> None: |
|
|
test_table_add_column = catalog.load_table("default.test_table_add_column") |
|
|
arrow_table = test_table_add_column.scan(row_filter="b == '2'").to_arrow() |
|
|
assert arrow_table["b"].to_pylist() == ["2"] |
|
|
|
|
|
arrow_table = test_table_add_column.scan(row_filter="b is not null").to_arrow() |
|
|
assert arrow_table["b"].to_pylist() == ["2"] |
|
|
|
|
|
arrow_table = test_table_add_column.scan(row_filter="b is null").to_arrow() |
|
|
assert arrow_table["b"].to_pylist() == [None] |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_upgrade_table_version(catalog: Catalog) -> None: |
|
|
table_test_table_version = catalog.load_table("default.test_table_version") |
|
|
|
|
|
assert table_test_table_version.format_version == 1 |
|
|
|
|
|
with table_test_table_version.transaction() as transaction: |
|
|
transaction.upgrade_table_version(format_version=1) |
|
|
|
|
|
assert table_test_table_version.format_version == 1 |
|
|
|
|
|
with table_test_table_version.transaction() as transaction: |
|
|
transaction.upgrade_table_version(format_version=2) |
|
|
|
|
|
assert table_test_table_version.format_version == 2 |
|
|
|
|
|
with pytest.raises(ValueError) as e: |
|
|
with table_test_table_version.transaction() as transaction: |
|
|
transaction.upgrade_table_version(format_version=1) |
|
|
assert "Cannot downgrade v2 table to v1" in str(e.value) |
|
|
|
|
|
with pytest.raises(ValueError) as e: |
|
|
with table_test_table_version.transaction() as transaction: |
|
|
transaction.upgrade_table_version(format_version=3) |
|
|
assert "Unsupported table format version: 3" in str(e.value) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_sanitize_character(catalog: Catalog) -> None: |
|
|
table_test_table_sanitized_character = catalog.load_table("default.test_table_sanitized_character") |
|
|
arrow_table = table_test_table_sanitized_character.scan().to_arrow() |
|
|
assert len(arrow_table.schema.names), 1 |
|
|
assert len(table_test_table_sanitized_character.schema().fields), 1 |
|
|
assert arrow_table.schema.names[0] == table_test_table_sanitized_character.schema().fields[0].name |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
@pytest.mark.parametrize("catalog", [pytest.lazy_fixture("session_catalog_hive"), pytest.lazy_fixture("session_catalog")]) |
|
|
def test_null_list_and_map(catalog: Catalog) -> None: |
|
|
table_test_empty_list_and_map = catalog.load_table("default.test_table_empty_list_and_map") |
|
|
arrow_table = table_test_empty_list_and_map.scan().to_arrow() |
|
|
assert arrow_table["col_list"].to_pylist() == [None, []] |
|
|
assert arrow_table["col_map"].to_pylist() == [None, []] |
|
|
|
|
|
|
|
|
|
|
|
assert arrow_table["col_list_with_struct"].to_pylist() == [[], [{"test": 1}]] |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
def test_hive_locking(session_catalog_hive: HiveCatalog) -> None: |
|
|
table = create_table(session_catalog_hive) |
|
|
|
|
|
database_name: str |
|
|
table_name: str |
|
|
_, database_name, table_name = table.identifier |
|
|
|
|
|
hive_client: _HiveClient = _HiveClient(session_catalog_hive.properties["uri"]) |
|
|
blocking_lock_request: LockRequest = session_catalog_hive._create_lock_request(database_name, table_name) |
|
|
|
|
|
with hive_client as open_client: |
|
|
|
|
|
lock: LockResponse = open_client.lock(blocking_lock_request) |
|
|
assert lock.state == LockState.ACQUIRED |
|
|
try: |
|
|
with pytest.raises(CommitFailedException, match="(Failed to acquire lock for).*"): |
|
|
table.transaction().set_properties(lock="fail").commit_transaction() |
|
|
finally: |
|
|
open_client.unlock(UnlockRequest(lock.lockid)) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
def test_hive_locking_with_retry(session_catalog_hive: HiveCatalog) -> None: |
|
|
table = create_table(session_catalog_hive) |
|
|
database_name: str |
|
|
table_name: str |
|
|
_, database_name, table_name = table.identifier |
|
|
session_catalog_hive._lock_check_min_wait_time = 0.1 |
|
|
session_catalog_hive._lock_check_max_wait_time = 0.5 |
|
|
session_catalog_hive._lock_check_retries = 5 |
|
|
|
|
|
hive_client: _HiveClient = _HiveClient(session_catalog_hive.properties["uri"]) |
|
|
|
|
|
executor = ExecutorFactory.get_or_create() |
|
|
|
|
|
with hive_client as open_client: |
|
|
|
|
|
def another_task() -> None: |
|
|
lock: LockResponse = open_client.lock(session_catalog_hive._create_lock_request(database_name, table_name)) |
|
|
time.sleep(1) |
|
|
open_client.unlock(UnlockRequest(lock.lockid)) |
|
|
|
|
|
|
|
|
executor.submit(another_task) |
|
|
time.sleep(0.5) |
|
|
|
|
|
table.transaction().set_properties(lock="xxx").commit_transaction() |
|
|
assert table.properties.get("lock") == "xxx" |
|
|
|