|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from datetime import date |
|
|
from typing import Iterator, Optional |
|
|
|
|
|
import pyarrow as pa |
|
|
import pyarrow.parquet as pq |
|
|
import pytest |
|
|
from pyspark.sql import SparkSession |
|
|
|
|
|
from pyiceberg.catalog import Catalog |
|
|
from pyiceberg.exceptions import NoSuchTableError |
|
|
from pyiceberg.partitioning import PartitionField, PartitionSpec |
|
|
from pyiceberg.schema import Schema |
|
|
from pyiceberg.table import Table |
|
|
from pyiceberg.transforms import BucketTransform, IdentityTransform, MonthTransform |
|
|
from pyiceberg.types import ( |
|
|
BooleanType, |
|
|
DateType, |
|
|
IntegerType, |
|
|
NestedField, |
|
|
StringType, |
|
|
) |
|
|
|
|
|
TABLE_SCHEMA = Schema( |
|
|
NestedField(field_id=1, name="foo", field_type=BooleanType(), required=False), |
|
|
NestedField(field_id=2, name="bar", field_type=StringType(), required=False), |
|
|
NestedField(field_id=4, name="baz", field_type=IntegerType(), required=False), |
|
|
NestedField(field_id=10, name="qux", field_type=DateType(), required=False), |
|
|
) |
|
|
|
|
|
ARROW_SCHEMA = pa.schema([ |
|
|
("foo", pa.bool_()), |
|
|
("bar", pa.string()), |
|
|
("baz", pa.int32()), |
|
|
("qux", pa.date32()), |
|
|
]) |
|
|
|
|
|
ARROW_TABLE = pa.Table.from_pylist( |
|
|
[ |
|
|
{ |
|
|
"foo": True, |
|
|
"bar": "bar_string", |
|
|
"baz": 123, |
|
|
"qux": date(2024, 3, 7), |
|
|
} |
|
|
], |
|
|
schema=ARROW_SCHEMA, |
|
|
) |
|
|
|
|
|
ARROW_SCHEMA_WITH_IDS = pa.schema([ |
|
|
pa.field("foo", pa.bool_(), nullable=False, metadata={"PARQUET:field_id": "1"}), |
|
|
pa.field("bar", pa.string(), nullable=False, metadata={"PARQUET:field_id": "2"}), |
|
|
pa.field("baz", pa.int32(), nullable=False, metadata={"PARQUET:field_id": "3"}), |
|
|
pa.field("qux", pa.date32(), nullable=False, metadata={"PARQUET:field_id": "4"}), |
|
|
]) |
|
|
|
|
|
|
|
|
ARROW_TABLE_WITH_IDS = pa.Table.from_pylist( |
|
|
[ |
|
|
{ |
|
|
"foo": True, |
|
|
"bar": "bar_string", |
|
|
"baz": 123, |
|
|
"qux": date(2024, 3, 7), |
|
|
} |
|
|
], |
|
|
schema=ARROW_SCHEMA_WITH_IDS, |
|
|
) |
|
|
|
|
|
ARROW_SCHEMA_UPDATED = pa.schema([ |
|
|
("foo", pa.bool_()), |
|
|
("baz", pa.int32()), |
|
|
("qux", pa.date32()), |
|
|
("quux", pa.int32()), |
|
|
]) |
|
|
|
|
|
ARROW_TABLE_UPDATED = pa.Table.from_pylist( |
|
|
[ |
|
|
{ |
|
|
"foo": True, |
|
|
"baz": 123, |
|
|
"qux": date(2024, 3, 7), |
|
|
"quux": 234, |
|
|
} |
|
|
], |
|
|
schema=ARROW_SCHEMA_UPDATED, |
|
|
) |
|
|
|
|
|
|
|
|
def _create_table( |
|
|
session_catalog: Catalog, identifier: str, format_version: int, partition_spec: Optional[PartitionSpec] = None |
|
|
) -> Table: |
|
|
try: |
|
|
session_catalog.drop_table(identifier=identifier) |
|
|
except NoSuchTableError: |
|
|
pass |
|
|
|
|
|
tbl = session_catalog.create_table( |
|
|
identifier=identifier, |
|
|
schema=TABLE_SCHEMA, |
|
|
properties={"format-version": str(format_version)}, |
|
|
partition_spec=partition_spec if partition_spec else PartitionSpec(), |
|
|
) |
|
|
|
|
|
return tbl |
|
|
|
|
|
|
|
|
@pytest.fixture(name="format_version", params=[pytest.param(1, id="format_version=1"), pytest.param(2, id="format_version=2")]) |
|
|
def format_version_fixure(request: pytest.FixtureRequest) -> Iterator[int]: |
|
|
"""Fixture to run tests with different table format versions.""" |
|
|
yield request.param |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
def test_add_files_to_unpartitioned_table(spark: SparkSession, session_catalog: Catalog, format_version: int) -> None: |
|
|
identifier = f"default.unpartitioned_table_v{format_version}" |
|
|
tbl = _create_table(session_catalog, identifier, format_version) |
|
|
|
|
|
file_paths = [f"s3://warehouse/default/unpartitioned/v{format_version}/test-{i}.parquet" for i in range(5)] |
|
|
|
|
|
for file_path in file_paths: |
|
|
fo = tbl.io.new_output(file_path) |
|
|
with fo.create(overwrite=True) as fos: |
|
|
with pq.ParquetWriter(fos, schema=ARROW_SCHEMA) as writer: |
|
|
writer.write_table(ARROW_TABLE) |
|
|
|
|
|
|
|
|
tbl.add_files(file_paths=file_paths) |
|
|
|
|
|
|
|
|
assert tbl.name_mapping() is not None |
|
|
|
|
|
rows = spark.sql( |
|
|
f""" |
|
|
SELECT added_data_files_count, existing_data_files_count, deleted_data_files_count |
|
|
FROM {identifier}.all_manifests |
|
|
""" |
|
|
).collect() |
|
|
|
|
|
assert [row.added_data_files_count for row in rows] == [5] |
|
|
assert [row.existing_data_files_count for row in rows] == [0] |
|
|
assert [row.deleted_data_files_count for row in rows] == [0] |
|
|
|
|
|
df = spark.table(identifier) |
|
|
assert df.count() == 5, "Expected 5 rows" |
|
|
for col in df.columns: |
|
|
assert df.filter(df[col].isNotNull()).count() == 5, "Expected all 5 rows to be non-null" |
|
|
|
|
|
|
|
|
assert len(tbl.scan().to_arrow()) == 5, "Expected 5 rows" |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
def test_add_files_to_unpartitioned_table_raises_file_not_found( |
|
|
spark: SparkSession, session_catalog: Catalog, format_version: int |
|
|
) -> None: |
|
|
identifier = f"default.unpartitioned_raises_not_found_v{format_version}" |
|
|
tbl = _create_table(session_catalog, identifier, format_version) |
|
|
|
|
|
file_paths = [f"s3://warehouse/default/unpartitioned_raises_not_found/v{format_version}/test-{i}.parquet" for i in range(5)] |
|
|
|
|
|
for file_path in file_paths: |
|
|
fo = tbl.io.new_output(file_path) |
|
|
with fo.create(overwrite=True) as fos: |
|
|
with pq.ParquetWriter(fos, schema=ARROW_SCHEMA) as writer: |
|
|
writer.write_table(ARROW_TABLE) |
|
|
|
|
|
|
|
|
with pytest.raises(FileNotFoundError): |
|
|
tbl.add_files(file_paths=file_paths + ["s3://warehouse/default/unpartitioned_raises_not_found/unknown.parquet"]) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
def test_add_files_to_unpartitioned_table_raises_has_field_ids( |
|
|
spark: SparkSession, session_catalog: Catalog, format_version: int |
|
|
) -> None: |
|
|
identifier = f"default.unpartitioned_raises_field_ids_v{format_version}" |
|
|
tbl = _create_table(session_catalog, identifier, format_version) |
|
|
|
|
|
file_paths = [f"s3://warehouse/default/unpartitioned_raises_field_ids/v{format_version}/test-{i}.parquet" for i in range(5)] |
|
|
|
|
|
for file_path in file_paths: |
|
|
fo = tbl.io.new_output(file_path) |
|
|
with fo.create(overwrite=True) as fos: |
|
|
with pq.ParquetWriter(fos, schema=ARROW_SCHEMA_WITH_IDS) as writer: |
|
|
writer.write_table(ARROW_TABLE_WITH_IDS) |
|
|
|
|
|
|
|
|
with pytest.raises(NotImplementedError): |
|
|
tbl.add_files(file_paths=file_paths) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
def test_add_files_to_unpartitioned_table_with_schema_updates( |
|
|
spark: SparkSession, session_catalog: Catalog, format_version: int |
|
|
) -> None: |
|
|
identifier = f"default.unpartitioned_table_schema_updates_v{format_version}" |
|
|
tbl = _create_table(session_catalog, identifier, format_version) |
|
|
|
|
|
file_paths = [f"s3://warehouse/default/unpartitioned_schema_updates/v{format_version}/test-{i}.parquet" for i in range(5)] |
|
|
|
|
|
for file_path in file_paths: |
|
|
fo = tbl.io.new_output(file_path) |
|
|
with fo.create(overwrite=True) as fos: |
|
|
with pq.ParquetWriter(fos, schema=ARROW_SCHEMA) as writer: |
|
|
writer.write_table(ARROW_TABLE) |
|
|
|
|
|
|
|
|
tbl.add_files(file_paths=file_paths) |
|
|
|
|
|
|
|
|
assert tbl.name_mapping() is not None |
|
|
|
|
|
with tbl.update_schema() as update: |
|
|
update.add_column("quux", IntegerType()) |
|
|
update.delete_column("bar") |
|
|
|
|
|
file_path = f"s3://warehouse/default/unpartitioned_schema_updates/v{format_version}/test-6.parquet" |
|
|
|
|
|
fo = tbl.io.new_output(file_path) |
|
|
with fo.create(overwrite=True) as fos: |
|
|
with pq.ParquetWriter(fos, schema=ARROW_SCHEMA_UPDATED) as writer: |
|
|
writer.write_table(ARROW_TABLE_UPDATED) |
|
|
|
|
|
|
|
|
tbl.add_files(file_paths=[file_path]) |
|
|
rows = spark.sql( |
|
|
f""" |
|
|
SELECT added_data_files_count, existing_data_files_count, deleted_data_files_count |
|
|
FROM {identifier}.all_manifests |
|
|
""" |
|
|
).collect() |
|
|
|
|
|
assert [row.added_data_files_count for row in rows] == [5, 1, 5] |
|
|
assert [row.existing_data_files_count for row in rows] == [0, 0, 0] |
|
|
assert [row.deleted_data_files_count for row in rows] == [0, 0, 0] |
|
|
|
|
|
df = spark.table(identifier) |
|
|
assert df.count() == 6, "Expected 6 rows" |
|
|
assert len(df.columns) == 4, "Expected 4 columns" |
|
|
|
|
|
for col in df.columns: |
|
|
value_count = 1 if col == "quux" else 6 |
|
|
assert df.filter(df[col].isNotNull()).count() == value_count, f"Expected {value_count} rows to be non-null" |
|
|
|
|
|
|
|
|
assert len(tbl.scan().to_arrow()) == 6, "Expected 6 rows" |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
def test_add_files_to_partitioned_table(spark: SparkSession, session_catalog: Catalog, format_version: int) -> None: |
|
|
identifier = f"default.partitioned_table_v{format_version}" |
|
|
|
|
|
partition_spec = PartitionSpec( |
|
|
PartitionField(source_id=4, field_id=1000, transform=IdentityTransform(), name="baz"), |
|
|
PartitionField(source_id=10, field_id=1001, transform=MonthTransform(), name="qux_month"), |
|
|
spec_id=0, |
|
|
) |
|
|
|
|
|
tbl = _create_table(session_catalog, identifier, format_version, partition_spec) |
|
|
|
|
|
date_iter = iter([date(2024, 3, 7), date(2024, 3, 8), date(2024, 3, 16), date(2024, 3, 18), date(2024, 3, 19)]) |
|
|
|
|
|
file_paths = [f"s3://warehouse/default/partitioned/v{format_version}/test-{i}.parquet" for i in range(5)] |
|
|
|
|
|
for file_path in file_paths: |
|
|
fo = tbl.io.new_output(file_path) |
|
|
with fo.create(overwrite=True) as fos: |
|
|
with pq.ParquetWriter(fos, schema=ARROW_SCHEMA) as writer: |
|
|
writer.write_table( |
|
|
pa.Table.from_pylist( |
|
|
[ |
|
|
{ |
|
|
"foo": True, |
|
|
"bar": "bar_string", |
|
|
"baz": 123, |
|
|
"qux": next(date_iter), |
|
|
} |
|
|
], |
|
|
schema=ARROW_SCHEMA, |
|
|
) |
|
|
) |
|
|
|
|
|
|
|
|
tbl.add_files(file_paths=file_paths) |
|
|
|
|
|
|
|
|
assert tbl.name_mapping() is not None |
|
|
|
|
|
rows = spark.sql( |
|
|
f""" |
|
|
SELECT added_data_files_count, existing_data_files_count, deleted_data_files_count |
|
|
FROM {identifier}.all_manifests |
|
|
""" |
|
|
).collect() |
|
|
|
|
|
assert [row.added_data_files_count for row in rows] == [5] |
|
|
assert [row.existing_data_files_count for row in rows] == [0] |
|
|
assert [row.deleted_data_files_count for row in rows] == [0] |
|
|
|
|
|
df = spark.table(identifier) |
|
|
assert df.count() == 5, "Expected 5 rows" |
|
|
for col in df.columns: |
|
|
assert df.filter(df[col].isNotNull()).count() == 5, "Expected all 5 rows to be non-null" |
|
|
|
|
|
partition_rows = spark.sql( |
|
|
f""" |
|
|
SELECT partition, record_count, file_count |
|
|
FROM {identifier}.partitions |
|
|
""" |
|
|
).collect() |
|
|
|
|
|
assert [row.record_count for row in partition_rows] == [5] |
|
|
assert [row.file_count for row in partition_rows] == [5] |
|
|
assert [(row.partition.baz, row.partition.qux_month) for row in partition_rows] == [(123, 650)] |
|
|
|
|
|
|
|
|
assert len(tbl.scan().to_arrow()) == 5, "Expected 5 rows" |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
def test_add_files_to_bucket_partitioned_table_fails(spark: SparkSession, session_catalog: Catalog, format_version: int) -> None: |
|
|
identifier = f"default.partitioned_table_bucket_fails_v{format_version}" |
|
|
|
|
|
partition_spec = PartitionSpec( |
|
|
PartitionField(source_id=4, field_id=1000, transform=BucketTransform(num_buckets=3), name="baz_bucket_3"), |
|
|
spec_id=0, |
|
|
) |
|
|
|
|
|
tbl = _create_table(session_catalog, identifier, format_version, partition_spec) |
|
|
|
|
|
int_iter = iter(range(5)) |
|
|
|
|
|
file_paths = [f"s3://warehouse/default/partitioned_table_bucket_fails/v{format_version}/test-{i}.parquet" for i in range(5)] |
|
|
|
|
|
for file_path in file_paths: |
|
|
fo = tbl.io.new_output(file_path) |
|
|
with fo.create(overwrite=True) as fos: |
|
|
with pq.ParquetWriter(fos, schema=ARROW_SCHEMA) as writer: |
|
|
writer.write_table( |
|
|
pa.Table.from_pylist( |
|
|
[ |
|
|
{ |
|
|
"foo": True, |
|
|
"bar": "bar_string", |
|
|
"baz": next(int_iter), |
|
|
"qux": date(2024, 3, 7), |
|
|
} |
|
|
], |
|
|
schema=ARROW_SCHEMA, |
|
|
) |
|
|
) |
|
|
|
|
|
|
|
|
with pytest.raises(ValueError) as exc_info: |
|
|
tbl.add_files(file_paths=file_paths) |
|
|
assert ( |
|
|
"Cannot infer partition value from parquet metadata for a non-linear Partition Field: baz_bucket_3 with transform bucket[3]" |
|
|
in str(exc_info.value) |
|
|
) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
def test_add_files_to_partitioned_table_fails_with_lower_and_upper_mismatch( |
|
|
spark: SparkSession, session_catalog: Catalog, format_version: int |
|
|
) -> None: |
|
|
identifier = f"default.partitioned_table_mismatch_fails_v{format_version}" |
|
|
|
|
|
partition_spec = PartitionSpec( |
|
|
PartitionField(source_id=4, field_id=1000, transform=IdentityTransform(), name="baz"), |
|
|
spec_id=0, |
|
|
) |
|
|
|
|
|
tbl = _create_table(session_catalog, identifier, format_version, partition_spec) |
|
|
|
|
|
file_paths = [f"s3://warehouse/default/partitioned_table_mismatch_fails/v{format_version}/test-{i}.parquet" for i in range(5)] |
|
|
|
|
|
for file_path in file_paths: |
|
|
fo = tbl.io.new_output(file_path) |
|
|
with fo.create(overwrite=True) as fos: |
|
|
with pq.ParquetWriter(fos, schema=ARROW_SCHEMA) as writer: |
|
|
writer.write_table( |
|
|
pa.Table.from_pylist( |
|
|
[ |
|
|
{ |
|
|
"foo": True, |
|
|
"bar": "bar_string", |
|
|
"baz": 123, |
|
|
"qux": date(2024, 3, 7), |
|
|
}, |
|
|
{ |
|
|
"foo": True, |
|
|
"bar": "bar_string", |
|
|
"baz": 124, |
|
|
"qux": date(2024, 3, 7), |
|
|
}, |
|
|
], |
|
|
schema=ARROW_SCHEMA, |
|
|
) |
|
|
) |
|
|
|
|
|
|
|
|
with pytest.raises(ValueError) as exc_info: |
|
|
tbl.add_files(file_paths=file_paths) |
|
|
assert ( |
|
|
"Cannot infer partition value from parquet metadata as there are more than one partition values for Partition Field: baz. lower_value=123, upper_value=124" |
|
|
in str(exc_info.value) |
|
|
) |
|
|
|
|
|
|
|
|
@pytest.mark.integration |
|
|
def test_add_files_snapshot_properties(spark: SparkSession, session_catalog: Catalog, format_version: int) -> None: |
|
|
identifier = f"default.unpartitioned_table_v{format_version}" |
|
|
tbl = _create_table(session_catalog, identifier, format_version) |
|
|
|
|
|
file_paths = [f"s3://warehouse/default/unpartitioned/v{format_version}/test-{i}.parquet" for i in range(5)] |
|
|
|
|
|
for file_path in file_paths: |
|
|
fo = tbl.io.new_output(file_path) |
|
|
with fo.create(overwrite=True) as fos: |
|
|
with pq.ParquetWriter(fos, schema=ARROW_SCHEMA) as writer: |
|
|
writer.write_table(ARROW_TABLE) |
|
|
|
|
|
|
|
|
tbl.add_files(file_paths=file_paths, snapshot_properties={"snapshot_prop_a": "test_prop_a"}) |
|
|
|
|
|
|
|
|
assert tbl.name_mapping() is not None |
|
|
|
|
|
summary = spark.sql(f"SELECT * FROM {identifier}.snapshots;").collect()[0].summary |
|
|
|
|
|
assert "snapshot_prop_a" in summary |
|
|
assert summary["snapshot_prop_a"] == "test_prop_a" |
|
|
|