|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import math |
|
|
from abc import ABC, abstractmethod |
|
|
from copy import copy |
|
|
from enum import Enum |
|
|
from types import TracebackType |
|
|
from typing import ( |
|
|
Any, |
|
|
Dict, |
|
|
Iterator, |
|
|
List, |
|
|
Literal, |
|
|
Optional, |
|
|
Type, |
|
|
) |
|
|
|
|
|
from pyiceberg.avro.file import AvroFile, AvroOutputFile |
|
|
from pyiceberg.conversions import to_bytes |
|
|
from pyiceberg.exceptions import ValidationError |
|
|
from pyiceberg.io import FileIO, InputFile, OutputFile |
|
|
from pyiceberg.partitioning import PartitionSpec |
|
|
from pyiceberg.schema import Schema |
|
|
from pyiceberg.typedef import EMPTY_DICT, Record, TableVersion |
|
|
from pyiceberg.types import ( |
|
|
BinaryType, |
|
|
BooleanType, |
|
|
IntegerType, |
|
|
ListType, |
|
|
LongType, |
|
|
MapType, |
|
|
NestedField, |
|
|
PrimitiveType, |
|
|
StringType, |
|
|
StructType, |
|
|
) |
|
|
|
|
|
UNASSIGNED_SEQ = -1 |
|
|
DEFAULT_BLOCK_SIZE = 67108864 |
|
|
DEFAULT_READ_VERSION: Literal[2] = 2 |
|
|
|
|
|
INITIAL_SEQUENCE_NUMBER = 0 |
|
|
|
|
|
|
|
|
class DataFileContent(int, Enum): |
|
|
DATA = 0 |
|
|
POSITION_DELETES = 1 |
|
|
EQUALITY_DELETES = 2 |
|
|
|
|
|
def __repr__(self) -> str: |
|
|
"""Return the string representation of the DataFileContent class.""" |
|
|
return f"DataFileContent.{self.name}" |
|
|
|
|
|
|
|
|
class ManifestContent(int, Enum): |
|
|
DATA = 0 |
|
|
DELETES = 1 |
|
|
|
|
|
def __repr__(self) -> str: |
|
|
"""Return the string representation of the ManifestContent class.""" |
|
|
return f"ManifestContent.{self.name}" |
|
|
|
|
|
|
|
|
class ManifestEntryStatus(int, Enum): |
|
|
EXISTING = 0 |
|
|
ADDED = 1 |
|
|
DELETED = 2 |
|
|
|
|
|
def __repr__(self) -> str: |
|
|
"""Return the string representation of the ManifestEntryStatus class.""" |
|
|
return f"ManifestEntryStatus.{self.name}" |
|
|
|
|
|
|
|
|
class FileFormat(str, Enum): |
|
|
AVRO = "AVRO" |
|
|
PARQUET = "PARQUET" |
|
|
ORC = "ORC" |
|
|
|
|
|
def __repr__(self) -> str: |
|
|
"""Return the string representation of the FileFormat class.""" |
|
|
return f"FileFormat.{self.name}" |
|
|
|
|
|
|
|
|
DATA_FILE_TYPE: Dict[int, StructType] = { |
|
|
1: StructType( |
|
|
NestedField(field_id=100, name="file_path", field_type=StringType(), required=True, doc="Location URI with FS scheme"), |
|
|
NestedField( |
|
|
field_id=101, |
|
|
name="file_format", |
|
|
field_type=StringType(), |
|
|
required=True, |
|
|
doc="File format name: avro, orc, or parquet", |
|
|
), |
|
|
NestedField( |
|
|
field_id=102, |
|
|
name="partition", |
|
|
field_type=StructType(), |
|
|
required=True, |
|
|
doc="Partition data tuple, schema based on the partition spec", |
|
|
), |
|
|
NestedField(field_id=103, name="record_count", field_type=LongType(), required=True, doc="Number of records in the file"), |
|
|
NestedField( |
|
|
field_id=104, name="file_size_in_bytes", field_type=LongType(), required=True, doc="Total file size in bytes" |
|
|
), |
|
|
NestedField( |
|
|
field_id=105, |
|
|
name="block_size_in_bytes", |
|
|
field_type=LongType(), |
|
|
required=True, |
|
|
doc="Deprecated. Always write a default in v1. Do not write in v2.", |
|
|
write_default=DEFAULT_BLOCK_SIZE, |
|
|
), |
|
|
NestedField( |
|
|
field_id=108, |
|
|
name="column_sizes", |
|
|
field_type=MapType(key_id=117, key_type=IntegerType(), value_id=118, value_type=LongType()), |
|
|
required=False, |
|
|
doc="Map of column id to total size on disk", |
|
|
), |
|
|
NestedField( |
|
|
field_id=109, |
|
|
name="value_counts", |
|
|
field_type=MapType(key_id=119, key_type=IntegerType(), value_id=120, value_type=LongType()), |
|
|
required=False, |
|
|
doc="Map of column id to total count, including null and NaN", |
|
|
), |
|
|
NestedField( |
|
|
field_id=110, |
|
|
name="null_value_counts", |
|
|
field_type=MapType(key_id=121, key_type=IntegerType(), value_id=122, value_type=LongType()), |
|
|
required=False, |
|
|
doc="Map of column id to null value count", |
|
|
), |
|
|
NestedField( |
|
|
field_id=137, |
|
|
name="nan_value_counts", |
|
|
field_type=MapType(key_id=138, key_type=IntegerType(), value_id=139, value_type=LongType()), |
|
|
required=False, |
|
|
doc="Map of column id to number of NaN values in the column", |
|
|
), |
|
|
NestedField( |
|
|
field_id=125, |
|
|
name="lower_bounds", |
|
|
field_type=MapType(key_id=126, key_type=IntegerType(), value_id=127, value_type=BinaryType()), |
|
|
required=False, |
|
|
doc="Map of column id to lower bound", |
|
|
), |
|
|
NestedField( |
|
|
field_id=128, |
|
|
name="upper_bounds", |
|
|
field_type=MapType(key_id=129, key_type=IntegerType(), value_id=130, value_type=BinaryType()), |
|
|
required=False, |
|
|
doc="Map of column id to upper bound", |
|
|
), |
|
|
NestedField( |
|
|
field_id=131, name="key_metadata", field_type=BinaryType(), required=False, doc="Encryption key metadata blob" |
|
|
), |
|
|
NestedField( |
|
|
field_id=132, |
|
|
name="split_offsets", |
|
|
field_type=ListType(element_id=133, element_type=LongType(), element_required=True), |
|
|
required=False, |
|
|
doc="Splittable offsets", |
|
|
), |
|
|
NestedField(field_id=140, name="sort_order_id", field_type=IntegerType(), required=False, doc="Sort order ID"), |
|
|
), |
|
|
2: StructType( |
|
|
NestedField( |
|
|
field_id=134, |
|
|
name="content", |
|
|
field_type=IntegerType(), |
|
|
required=True, |
|
|
doc="File format name: avro, orc, or parquet", |
|
|
initial_default=DataFileContent.DATA, |
|
|
), |
|
|
NestedField(field_id=100, name="file_path", field_type=StringType(), required=True, doc="Location URI with FS scheme"), |
|
|
NestedField( |
|
|
field_id=101, |
|
|
name="file_format", |
|
|
field_type=StringType(), |
|
|
required=True, |
|
|
doc="File format name: avro, orc, or parquet", |
|
|
), |
|
|
NestedField( |
|
|
field_id=102, |
|
|
name="partition", |
|
|
field_type=StructType(), |
|
|
required=True, |
|
|
doc="Partition data tuple, schema based on the partition spec", |
|
|
), |
|
|
NestedField(field_id=103, name="record_count", field_type=LongType(), required=True, doc="Number of records in the file"), |
|
|
NestedField( |
|
|
field_id=104, name="file_size_in_bytes", field_type=LongType(), required=True, doc="Total file size in bytes" |
|
|
), |
|
|
NestedField( |
|
|
field_id=108, |
|
|
name="column_sizes", |
|
|
field_type=MapType(key_id=117, key_type=IntegerType(), value_id=118, value_type=LongType()), |
|
|
required=False, |
|
|
doc="Map of column id to total size on disk", |
|
|
), |
|
|
NestedField( |
|
|
field_id=109, |
|
|
name="value_counts", |
|
|
field_type=MapType(key_id=119, key_type=IntegerType(), value_id=120, value_type=LongType()), |
|
|
required=False, |
|
|
doc="Map of column id to total count, including null and NaN", |
|
|
), |
|
|
NestedField( |
|
|
field_id=110, |
|
|
name="null_value_counts", |
|
|
field_type=MapType(key_id=121, key_type=IntegerType(), value_id=122, value_type=LongType()), |
|
|
required=False, |
|
|
doc="Map of column id to null value count", |
|
|
), |
|
|
NestedField( |
|
|
field_id=137, |
|
|
name="nan_value_counts", |
|
|
field_type=MapType(key_id=138, key_type=IntegerType(), value_id=139, value_type=LongType()), |
|
|
required=False, |
|
|
doc="Map of column id to number of NaN values in the column", |
|
|
), |
|
|
NestedField( |
|
|
field_id=125, |
|
|
name="lower_bounds", |
|
|
field_type=MapType(key_id=126, key_type=IntegerType(), value_id=127, value_type=BinaryType()), |
|
|
required=False, |
|
|
doc="Map of column id to lower bound", |
|
|
), |
|
|
NestedField( |
|
|
field_id=128, |
|
|
name="upper_bounds", |
|
|
field_type=MapType(key_id=129, key_type=IntegerType(), value_id=130, value_type=BinaryType()), |
|
|
required=False, |
|
|
doc="Map of column id to upper bound", |
|
|
), |
|
|
NestedField( |
|
|
field_id=131, name="key_metadata", field_type=BinaryType(), required=False, doc="Encryption key metadata blob" |
|
|
), |
|
|
NestedField( |
|
|
field_id=132, |
|
|
name="split_offsets", |
|
|
field_type=ListType(element_id=133, element_type=LongType(), element_required=True), |
|
|
required=False, |
|
|
doc="Splittable offsets", |
|
|
), |
|
|
NestedField( |
|
|
field_id=135, |
|
|
name="equality_ids", |
|
|
field_type=ListType(element_id=136, element_type=LongType(), element_required=True), |
|
|
required=False, |
|
|
doc="Field ids used to determine row equality in equality delete files.", |
|
|
), |
|
|
NestedField( |
|
|
field_id=140, |
|
|
name="sort_order_id", |
|
|
field_type=IntegerType(), |
|
|
required=False, |
|
|
doc="ID representing sort order for this file", |
|
|
), |
|
|
), |
|
|
} |
|
|
|
|
|
|
|
|
def data_file_with_partition(partition_type: StructType, format_version: TableVersion) -> StructType: |
|
|
data_file_partition_type = StructType(*[ |
|
|
NestedField( |
|
|
field_id=field.field_id, |
|
|
name=field.name, |
|
|
field_type=field.field_type, |
|
|
required=field.required, |
|
|
) |
|
|
for field in partition_type.fields |
|
|
]) |
|
|
|
|
|
return StructType(*[ |
|
|
NestedField( |
|
|
field_id=102, |
|
|
name="partition", |
|
|
field_type=data_file_partition_type, |
|
|
required=True, |
|
|
doc="Partition data tuple, schema based on the partition spec", |
|
|
) |
|
|
if field.field_id == 102 |
|
|
else field |
|
|
for field in DATA_FILE_TYPE[format_version].fields |
|
|
]) |
|
|
|
|
|
|
|
|
class DataFile(Record): |
|
|
__slots__ = ( |
|
|
"content", |
|
|
"file_path", |
|
|
"file_format", |
|
|
"partition", |
|
|
"record_count", |
|
|
"file_size_in_bytes", |
|
|
"column_sizes", |
|
|
"value_counts", |
|
|
"null_value_counts", |
|
|
"nan_value_counts", |
|
|
"lower_bounds", |
|
|
"upper_bounds", |
|
|
"key_metadata", |
|
|
"split_offsets", |
|
|
"equality_ids", |
|
|
"sort_order_id", |
|
|
"spec_id", |
|
|
) |
|
|
content: DataFileContent |
|
|
file_path: str |
|
|
file_format: FileFormat |
|
|
partition: Record |
|
|
record_count: int |
|
|
file_size_in_bytes: int |
|
|
column_sizes: Dict[int, int] |
|
|
value_counts: Dict[int, int] |
|
|
null_value_counts: Dict[int, int] |
|
|
nan_value_counts: Dict[int, int] |
|
|
lower_bounds: Dict[int, bytes] |
|
|
upper_bounds: Dict[int, bytes] |
|
|
key_metadata: Optional[bytes] |
|
|
split_offsets: Optional[List[int]] |
|
|
equality_ids: Optional[List[int]] |
|
|
sort_order_id: Optional[int] |
|
|
spec_id: Optional[int] |
|
|
|
|
|
def __setattr__(self, name: str, value: Any) -> None: |
|
|
"""Assign a key/value to a DataFile.""" |
|
|
|
|
|
if name == "file_format": |
|
|
value = FileFormat[value] |
|
|
super().__setattr__(name, value) |
|
|
|
|
|
def __init__(self, format_version: TableVersion = DEFAULT_READ_VERSION, *data: Any, **named_data: Any) -> None: |
|
|
super().__init__( |
|
|
*data, |
|
|
**{"struct": DATA_FILE_TYPE[format_version], **named_data}, |
|
|
) |
|
|
|
|
|
def __hash__(self) -> int: |
|
|
"""Return the hash of the file path.""" |
|
|
return hash(self.file_path) |
|
|
|
|
|
def __eq__(self, other: Any) -> bool: |
|
|
"""Compare the datafile with another object. |
|
|
|
|
|
If it is a datafile, it will compare based on the file_path. |
|
|
""" |
|
|
return self.file_path == other.file_path if isinstance(other, DataFile) else False |
|
|
|
|
|
|
|
|
MANIFEST_ENTRY_SCHEMAS = { |
|
|
1: Schema( |
|
|
NestedField(0, "status", IntegerType(), required=True), |
|
|
NestedField(1, "snapshot_id", LongType(), required=True), |
|
|
NestedField(2, "data_file", DATA_FILE_TYPE[1], required=True), |
|
|
), |
|
|
2: Schema( |
|
|
NestedField(0, "status", IntegerType(), required=True), |
|
|
NestedField(1, "snapshot_id", LongType(), required=False), |
|
|
NestedField(3, "data_sequence_number", LongType(), required=False), |
|
|
NestedField(4, "file_sequence_number", LongType(), required=False), |
|
|
NestedField(2, "data_file", DATA_FILE_TYPE[2], required=True), |
|
|
), |
|
|
} |
|
|
|
|
|
MANIFEST_ENTRY_SCHEMAS_STRUCT = {format_version: schema.as_struct() for format_version, schema in MANIFEST_ENTRY_SCHEMAS.items()} |
|
|
|
|
|
|
|
|
def manifest_entry_schema_with_data_file(format_version: TableVersion, data_file: StructType) -> Schema: |
|
|
return Schema(*[ |
|
|
NestedField(2, "data_file", data_file, required=True) if field.field_id == 2 else field |
|
|
for field in MANIFEST_ENTRY_SCHEMAS[format_version].fields |
|
|
]) |
|
|
|
|
|
|
|
|
class ManifestEntry(Record): |
|
|
__slots__ = ("status", "snapshot_id", "data_sequence_number", "file_sequence_number", "data_file") |
|
|
status: ManifestEntryStatus |
|
|
snapshot_id: Optional[int] |
|
|
data_sequence_number: Optional[int] |
|
|
file_sequence_number: Optional[int] |
|
|
data_file: DataFile |
|
|
|
|
|
def __init__(self, *data: Any, **named_data: Any) -> None: |
|
|
super().__init__(*data, **{"struct": MANIFEST_ENTRY_SCHEMAS_STRUCT[DEFAULT_READ_VERSION], **named_data}) |
|
|
|
|
|
|
|
|
PARTITION_FIELD_SUMMARY_TYPE = StructType( |
|
|
NestedField(509, "contains_null", BooleanType(), required=True), |
|
|
NestedField(518, "contains_nan", BooleanType(), required=False), |
|
|
NestedField(510, "lower_bound", BinaryType(), required=False), |
|
|
NestedField(511, "upper_bound", BinaryType(), required=False), |
|
|
) |
|
|
|
|
|
|
|
|
class PartitionFieldSummary(Record): |
|
|
__slots__ = ("contains_null", "contains_nan", "lower_bound", "upper_bound") |
|
|
contains_null: bool |
|
|
contains_nan: Optional[bool] |
|
|
lower_bound: Optional[bytes] |
|
|
upper_bound: Optional[bytes] |
|
|
|
|
|
def __init__(self, *data: Any, **named_data: Any) -> None: |
|
|
super().__init__(*data, **{"struct": PARTITION_FIELD_SUMMARY_TYPE, **named_data}) |
|
|
|
|
|
|
|
|
class PartitionFieldStats: |
|
|
_type: PrimitiveType |
|
|
_contains_null: bool |
|
|
_contains_nan: bool |
|
|
_min: Optional[Any] |
|
|
_max: Optional[Any] |
|
|
|
|
|
def __init__(self, iceberg_type: PrimitiveType) -> None: |
|
|
self._type = iceberg_type |
|
|
self._contains_null = False |
|
|
self._contains_nan = False |
|
|
self._min = None |
|
|
self._max = None |
|
|
|
|
|
def to_summary(self) -> PartitionFieldSummary: |
|
|
return PartitionFieldSummary( |
|
|
contains_null=self._contains_null, |
|
|
contains_nan=self._contains_nan, |
|
|
lower_bound=to_bytes(self._type, self._min) if self._min is not None else None, |
|
|
upper_bound=to_bytes(self._type, self._max) if self._max is not None else None, |
|
|
) |
|
|
|
|
|
def update(self, value: Any) -> None: |
|
|
if value is None: |
|
|
self._contains_null = True |
|
|
elif isinstance(value, float) and math.isnan(value): |
|
|
self._contains_nan = True |
|
|
else: |
|
|
if self._min is None: |
|
|
self._min = value |
|
|
self._max = value |
|
|
else: |
|
|
self._max = max(self._max, value) |
|
|
self._min = min(self._min, value) |
|
|
|
|
|
|
|
|
def construct_partition_summaries(spec: PartitionSpec, schema: Schema, partitions: List[Record]) -> List[PartitionFieldSummary]: |
|
|
types = [field.field_type for field in spec.partition_type(schema).fields] |
|
|
field_stats = [PartitionFieldStats(field_type) for field_type in types] |
|
|
for partition_keys in partitions: |
|
|
for i, field_type in enumerate(types): |
|
|
if not isinstance(field_type, PrimitiveType): |
|
|
raise ValueError(f"Expected a primitive type for the partition field, got {field_type}") |
|
|
partition_key = partition_keys[i] |
|
|
field_stats[i].update(partition_key) |
|
|
return [field.to_summary() for field in field_stats] |
|
|
|
|
|
|
|
|
MANIFEST_LIST_FILE_SCHEMAS: Dict[int, Schema] = { |
|
|
1: Schema( |
|
|
NestedField(500, "manifest_path", StringType(), required=True, doc="Location URI with FS scheme"), |
|
|
NestedField(501, "manifest_length", LongType(), required=True), |
|
|
NestedField(502, "partition_spec_id", IntegerType(), required=True), |
|
|
NestedField(503, "added_snapshot_id", LongType(), required=True), |
|
|
NestedField(504, "added_files_count", IntegerType(), required=False), |
|
|
NestedField(505, "existing_files_count", IntegerType(), required=False), |
|
|
NestedField(506, "deleted_files_count", IntegerType(), required=False), |
|
|
NestedField(512, "added_rows_count", LongType(), required=False), |
|
|
NestedField(513, "existing_rows_count", LongType(), required=False), |
|
|
NestedField(514, "deleted_rows_count", LongType(), required=False), |
|
|
NestedField(507, "partitions", ListType(508, PARTITION_FIELD_SUMMARY_TYPE, element_required=True), required=False), |
|
|
NestedField(519, "key_metadata", BinaryType(), required=False), |
|
|
), |
|
|
2: Schema( |
|
|
NestedField(500, "manifest_path", StringType(), required=True, doc="Location URI with FS scheme"), |
|
|
NestedField(501, "manifest_length", LongType(), required=True), |
|
|
NestedField(502, "partition_spec_id", IntegerType(), required=True), |
|
|
NestedField(517, "content", IntegerType(), required=True, initial_default=ManifestContent.DATA), |
|
|
NestedField(515, "sequence_number", LongType(), required=True, initial_default=0), |
|
|
NestedField(516, "min_sequence_number", LongType(), required=True, initial_default=0), |
|
|
NestedField(503, "added_snapshot_id", LongType(), required=True), |
|
|
NestedField(504, "added_files_count", IntegerType(), required=True), |
|
|
NestedField(505, "existing_files_count", IntegerType(), required=True), |
|
|
NestedField(506, "deleted_files_count", IntegerType(), required=True), |
|
|
NestedField(512, "added_rows_count", LongType(), required=True), |
|
|
NestedField(513, "existing_rows_count", LongType(), required=True), |
|
|
NestedField(514, "deleted_rows_count", LongType(), required=True), |
|
|
NestedField(507, "partitions", ListType(508, PARTITION_FIELD_SUMMARY_TYPE, element_required=True), required=False), |
|
|
NestedField(519, "key_metadata", BinaryType(), required=False), |
|
|
), |
|
|
} |
|
|
|
|
|
MANIFEST_LIST_FILE_STRUCTS = {format_version: schema.as_struct() for format_version, schema in MANIFEST_LIST_FILE_SCHEMAS.items()} |
|
|
|
|
|
|
|
|
POSITIONAL_DELETE_SCHEMA = Schema( |
|
|
NestedField(2147483546, "file_path", StringType()), NestedField(2147483545, "pos", IntegerType()) |
|
|
) |
|
|
|
|
|
|
|
|
class ManifestFile(Record): |
|
|
__slots__ = ( |
|
|
"manifest_path", |
|
|
"manifest_length", |
|
|
"partition_spec_id", |
|
|
"content", |
|
|
"sequence_number", |
|
|
"min_sequence_number", |
|
|
"added_snapshot_id", |
|
|
"added_files_count", |
|
|
"existing_files_count", |
|
|
"deleted_files_count", |
|
|
"added_rows_count", |
|
|
"existing_rows_count", |
|
|
"deleted_rows_count", |
|
|
"partitions", |
|
|
"key_metadata", |
|
|
) |
|
|
manifest_path: str |
|
|
manifest_length: int |
|
|
partition_spec_id: int |
|
|
content: ManifestContent |
|
|
sequence_number: int |
|
|
min_sequence_number: int |
|
|
added_snapshot_id: int |
|
|
added_files_count: Optional[int] |
|
|
existing_files_count: Optional[int] |
|
|
deleted_files_count: Optional[int] |
|
|
added_rows_count: Optional[int] |
|
|
existing_rows_count: Optional[int] |
|
|
deleted_rows_count: Optional[int] |
|
|
partitions: Optional[List[PartitionFieldSummary]] |
|
|
key_metadata: Optional[bytes] |
|
|
|
|
|
def __init__(self, *data: Any, **named_data: Any) -> None: |
|
|
super().__init__(*data, **{"struct": MANIFEST_LIST_FILE_STRUCTS[DEFAULT_READ_VERSION], **named_data}) |
|
|
|
|
|
def has_added_files(self) -> bool: |
|
|
return self.added_files_count is None or self.added_files_count > 0 |
|
|
|
|
|
def has_existing_files(self) -> bool: |
|
|
return self.existing_files_count is None or self.existing_files_count > 0 |
|
|
|
|
|
def fetch_manifest_entry(self, io: FileIO, discard_deleted: bool = True) -> List[ManifestEntry]: |
|
|
""" |
|
|
Read the manifest entries from the manifest file. |
|
|
|
|
|
Args: |
|
|
io: The FileIO to fetch the file. |
|
|
discard_deleted: Filter on live entries. |
|
|
|
|
|
Returns: |
|
|
An Iterator of manifest entries. |
|
|
""" |
|
|
input_file = io.new_input(self.manifest_path) |
|
|
with AvroFile[ManifestEntry]( |
|
|
input_file, |
|
|
MANIFEST_ENTRY_SCHEMAS[DEFAULT_READ_VERSION], |
|
|
read_types={-1: ManifestEntry, 2: DataFile}, |
|
|
read_enums={0: ManifestEntryStatus, 101: FileFormat, 134: DataFileContent}, |
|
|
) as reader: |
|
|
return [ |
|
|
_inherit_from_manifest(entry, self) |
|
|
for entry in reader |
|
|
if not discard_deleted or entry.status != ManifestEntryStatus.DELETED |
|
|
] |
|
|
|
|
|
|
|
|
def read_manifest_list(input_file: InputFile) -> Iterator[ManifestFile]: |
|
|
""" |
|
|
Read the manifests from the manifest list. |
|
|
|
|
|
Args: |
|
|
input_file: The input file where the stream can be read from. |
|
|
|
|
|
Returns: |
|
|
An iterator of ManifestFiles that are part of the list. |
|
|
""" |
|
|
with AvroFile[ManifestFile]( |
|
|
input_file, |
|
|
MANIFEST_LIST_FILE_SCHEMAS[DEFAULT_READ_VERSION], |
|
|
read_types={-1: ManifestFile, 508: PartitionFieldSummary}, |
|
|
read_enums={517: ManifestContent}, |
|
|
) as reader: |
|
|
yield from reader |
|
|
|
|
|
|
|
|
def _inherit_from_manifest(entry: ManifestEntry, manifest: ManifestFile) -> ManifestEntry: |
|
|
""" |
|
|
Inherits properties from manifest file. |
|
|
|
|
|
The properties that will be inherited are: |
|
|
- sequence numbers |
|
|
- partition spec id. |
|
|
|
|
|
More information about inheriting sequence numbers: https://iceberg.apache.org/spec/#sequence-number-inheritance |
|
|
|
|
|
Args: |
|
|
entry: The manifest entry. |
|
|
manifest: The manifest file. |
|
|
|
|
|
Returns: |
|
|
The manifest entry with properties inherited. |
|
|
""" |
|
|
|
|
|
|
|
|
if entry.snapshot_id is None: |
|
|
entry.snapshot_id = manifest.added_snapshot_id |
|
|
|
|
|
|
|
|
|
|
|
if entry.data_sequence_number is None and (manifest.sequence_number == 0 or entry.status == ManifestEntryStatus.ADDED): |
|
|
entry.data_sequence_number = manifest.sequence_number |
|
|
|
|
|
|
|
|
|
|
|
if entry.file_sequence_number is None and (manifest.sequence_number == 0 or entry.status == ManifestEntryStatus.ADDED): |
|
|
|
|
|
entry.file_sequence_number = manifest.sequence_number |
|
|
|
|
|
|
|
|
entry.data_file.spec_id = manifest.partition_spec_id |
|
|
|
|
|
return entry |
|
|
|
|
|
|
|
|
class ManifestWriter(ABC): |
|
|
closed: bool |
|
|
_spec: PartitionSpec |
|
|
_schema: Schema |
|
|
_output_file: OutputFile |
|
|
_writer: AvroOutputFile[ManifestEntry] |
|
|
_snapshot_id: int |
|
|
_meta: Dict[str, str] |
|
|
_added_files: int |
|
|
_added_rows: int |
|
|
_existing_files: int |
|
|
_existing_rows: int |
|
|
_deleted_files: int |
|
|
_deleted_rows: int |
|
|
_min_data_sequence_number: Optional[int] |
|
|
_partitions: List[Record] |
|
|
|
|
|
def __init__( |
|
|
self, spec: PartitionSpec, schema: Schema, output_file: OutputFile, snapshot_id: int, meta: Dict[str, str] = EMPTY_DICT |
|
|
) -> None: |
|
|
self.closed = False |
|
|
self._spec = spec |
|
|
self._schema = schema |
|
|
self._output_file = output_file |
|
|
self._snapshot_id = snapshot_id |
|
|
self._meta = meta |
|
|
|
|
|
self._added_files = 0 |
|
|
self._added_rows = 0 |
|
|
self._existing_files = 0 |
|
|
self._existing_rows = 0 |
|
|
self._deleted_files = 0 |
|
|
self._deleted_rows = 0 |
|
|
self._min_data_sequence_number = None |
|
|
self._partitions = [] |
|
|
|
|
|
def __enter__(self) -> ManifestWriter: |
|
|
"""Open the writer.""" |
|
|
self._writer = self.new_writer() |
|
|
self._writer.__enter__() |
|
|
return self |
|
|
|
|
|
def __exit__( |
|
|
self, |
|
|
exc_type: Optional[Type[BaseException]], |
|
|
exc_value: Optional[BaseException], |
|
|
traceback: Optional[TracebackType], |
|
|
) -> None: |
|
|
"""Close the writer.""" |
|
|
self.closed = True |
|
|
self._writer.__exit__(exc_type, exc_value, traceback) |
|
|
|
|
|
@abstractmethod |
|
|
def content(self) -> ManifestContent: ... |
|
|
|
|
|
@property |
|
|
@abstractmethod |
|
|
def version(self) -> TableVersion: ... |
|
|
|
|
|
def _with_partition(self, format_version: TableVersion) -> Schema: |
|
|
data_file_type = data_file_with_partition( |
|
|
format_version=format_version, partition_type=self._spec.partition_type(self._schema) |
|
|
) |
|
|
return manifest_entry_schema_with_data_file(format_version=format_version, data_file=data_file_type) |
|
|
|
|
|
def new_writer(self) -> AvroOutputFile[ManifestEntry]: |
|
|
return AvroOutputFile[ManifestEntry]( |
|
|
output_file=self._output_file, |
|
|
file_schema=self._with_partition(self.version), |
|
|
record_schema=self._with_partition(DEFAULT_READ_VERSION), |
|
|
schema_name="manifest_entry", |
|
|
metadata=self._meta, |
|
|
) |
|
|
|
|
|
@abstractmethod |
|
|
def prepare_entry(self, entry: ManifestEntry) -> ManifestEntry: ... |
|
|
|
|
|
def to_manifest_file(self) -> ManifestFile: |
|
|
"""Return the manifest file.""" |
|
|
|
|
|
self.closed = True |
|
|
min_sequence_number = self._min_data_sequence_number or UNASSIGNED_SEQ |
|
|
return ManifestFile( |
|
|
manifest_path=self._output_file.location, |
|
|
manifest_length=len(self._writer.output_file), |
|
|
partition_spec_id=self._spec.spec_id, |
|
|
content=self.content(), |
|
|
sequence_number=UNASSIGNED_SEQ, |
|
|
min_sequence_number=min_sequence_number, |
|
|
added_snapshot_id=self._snapshot_id, |
|
|
added_files_count=self._added_files, |
|
|
existing_files_count=self._existing_files, |
|
|
deleted_files_count=self._deleted_files, |
|
|
added_rows_count=self._added_rows, |
|
|
existing_rows_count=self._existing_rows, |
|
|
deleted_rows_count=self._deleted_rows, |
|
|
partitions=construct_partition_summaries(self._spec, self._schema, self._partitions), |
|
|
key_metadata=None, |
|
|
) |
|
|
|
|
|
def add_entry(self, entry: ManifestEntry) -> ManifestWriter: |
|
|
if self.closed: |
|
|
raise RuntimeError("Cannot add entry to closed manifest writer") |
|
|
if entry.status == ManifestEntryStatus.ADDED: |
|
|
self._added_files += 1 |
|
|
self._added_rows += entry.data_file.record_count |
|
|
elif entry.status == ManifestEntryStatus.EXISTING: |
|
|
self._existing_files += 1 |
|
|
self._existing_rows += entry.data_file.record_count |
|
|
elif entry.status == ManifestEntryStatus.DELETED: |
|
|
self._deleted_files += 1 |
|
|
self._deleted_rows += entry.data_file.record_count |
|
|
|
|
|
self._partitions.append(entry.data_file.partition) |
|
|
|
|
|
if ( |
|
|
(entry.status == ManifestEntryStatus.ADDED or entry.status == ManifestEntryStatus.EXISTING) |
|
|
and entry.data_sequence_number is not None |
|
|
and (self._min_data_sequence_number is None or entry.data_sequence_number < self._min_data_sequence_number) |
|
|
): |
|
|
self._min_data_sequence_number = entry.data_sequence_number |
|
|
|
|
|
self._writer.write_block([self.prepare_entry(entry)]) |
|
|
return self |
|
|
|
|
|
|
|
|
class ManifestWriterV1(ManifestWriter): |
|
|
def __init__(self, spec: PartitionSpec, schema: Schema, output_file: OutputFile, snapshot_id: int): |
|
|
super().__init__( |
|
|
spec, |
|
|
schema, |
|
|
output_file, |
|
|
snapshot_id, |
|
|
{ |
|
|
"schema": schema.model_dump_json(), |
|
|
"partition-spec": spec.model_dump_json(), |
|
|
"partition-spec-id": str(spec.spec_id), |
|
|
"format-version": "1", |
|
|
}, |
|
|
) |
|
|
|
|
|
def content(self) -> ManifestContent: |
|
|
return ManifestContent.DATA |
|
|
|
|
|
@property |
|
|
def version(self) -> TableVersion: |
|
|
return 1 |
|
|
|
|
|
def prepare_entry(self, entry: ManifestEntry) -> ManifestEntry: |
|
|
return entry |
|
|
|
|
|
|
|
|
class ManifestWriterV2(ManifestWriter): |
|
|
def __init__(self, spec: PartitionSpec, schema: Schema, output_file: OutputFile, snapshot_id: int): |
|
|
super().__init__( |
|
|
spec, |
|
|
schema, |
|
|
output_file, |
|
|
snapshot_id, |
|
|
meta={ |
|
|
"schema": schema.model_dump_json(), |
|
|
"partition-spec": spec.model_dump_json(), |
|
|
"partition-spec-id": str(spec.spec_id), |
|
|
"format-version": "2", |
|
|
"content": "data", |
|
|
}, |
|
|
) |
|
|
|
|
|
def content(self) -> ManifestContent: |
|
|
return ManifestContent.DATA |
|
|
|
|
|
@property |
|
|
def version(self) -> TableVersion: |
|
|
return 2 |
|
|
|
|
|
def prepare_entry(self, entry: ManifestEntry) -> ManifestEntry: |
|
|
if entry.data_sequence_number is None: |
|
|
if entry.snapshot_id is not None and entry.snapshot_id != self._snapshot_id: |
|
|
raise ValueError(f"Found unassigned sequence number for an entry from snapshot: {entry.snapshot_id}") |
|
|
if entry.status != ManifestEntryStatus.ADDED: |
|
|
raise ValueError("Only entries with status ADDED can have null sequence number") |
|
|
return entry |
|
|
|
|
|
|
|
|
def write_manifest( |
|
|
format_version: TableVersion, spec: PartitionSpec, schema: Schema, output_file: OutputFile, snapshot_id: int |
|
|
) -> ManifestWriter: |
|
|
if format_version == 1: |
|
|
return ManifestWriterV1(spec, schema, output_file, snapshot_id) |
|
|
elif format_version == 2: |
|
|
return ManifestWriterV2(spec, schema, output_file, snapshot_id) |
|
|
else: |
|
|
raise ValueError(f"Cannot write manifest for table version: {format_version}") |
|
|
|
|
|
|
|
|
class ManifestListWriter(ABC): |
|
|
_format_version: TableVersion |
|
|
_output_file: OutputFile |
|
|
_meta: Dict[str, str] |
|
|
_manifest_files: List[ManifestFile] |
|
|
_commit_snapshot_id: int |
|
|
_writer: AvroOutputFile[ManifestFile] |
|
|
|
|
|
def __init__(self, format_version: TableVersion, output_file: OutputFile, meta: Dict[str, Any]): |
|
|
self._format_version = format_version |
|
|
self._output_file = output_file |
|
|
self._meta = meta |
|
|
self._manifest_files = [] |
|
|
|
|
|
def __enter__(self) -> ManifestListWriter: |
|
|
"""Open the writer for writing.""" |
|
|
self._writer = AvroOutputFile[ManifestFile]( |
|
|
output_file=self._output_file, |
|
|
record_schema=MANIFEST_LIST_FILE_SCHEMAS[DEFAULT_READ_VERSION], |
|
|
file_schema=MANIFEST_LIST_FILE_SCHEMAS[self._format_version], |
|
|
schema_name="manifest_file", |
|
|
metadata=self._meta, |
|
|
) |
|
|
self._writer.__enter__() |
|
|
return self |
|
|
|
|
|
def __exit__( |
|
|
self, |
|
|
exc_type: Optional[Type[BaseException]], |
|
|
exc_value: Optional[BaseException], |
|
|
traceback: Optional[TracebackType], |
|
|
) -> None: |
|
|
"""Close the writer.""" |
|
|
self._writer.__exit__(exc_type, exc_value, traceback) |
|
|
return |
|
|
|
|
|
@abstractmethod |
|
|
def prepare_manifest(self, manifest_file: ManifestFile) -> ManifestFile: ... |
|
|
|
|
|
def add_manifests(self, manifest_files: List[ManifestFile]) -> ManifestListWriter: |
|
|
self._writer.write_block([self.prepare_manifest(manifest_file) for manifest_file in manifest_files]) |
|
|
return self |
|
|
|
|
|
|
|
|
class ManifestListWriterV1(ManifestListWriter): |
|
|
def __init__(self, output_file: OutputFile, snapshot_id: int, parent_snapshot_id: Optional[int]): |
|
|
super().__init__( |
|
|
format_version=1, |
|
|
output_file=output_file, |
|
|
meta={"snapshot-id": str(snapshot_id), "parent-snapshot-id": str(parent_snapshot_id), "format-version": "1"}, |
|
|
) |
|
|
|
|
|
def prepare_manifest(self, manifest_file: ManifestFile) -> ManifestFile: |
|
|
if manifest_file.content != ManifestContent.DATA: |
|
|
raise ValidationError("Cannot store delete manifests in a v1 table") |
|
|
return manifest_file |
|
|
|
|
|
|
|
|
class ManifestListWriterV2(ManifestListWriter): |
|
|
_commit_snapshot_id: int |
|
|
_sequence_number: int |
|
|
|
|
|
def __init__(self, output_file: OutputFile, snapshot_id: int, parent_snapshot_id: Optional[int], sequence_number: int): |
|
|
super().__init__( |
|
|
format_version=2, |
|
|
output_file=output_file, |
|
|
meta={ |
|
|
"snapshot-id": str(snapshot_id), |
|
|
"parent-snapshot-id": str(parent_snapshot_id), |
|
|
"sequence-number": str(sequence_number), |
|
|
"format-version": "2", |
|
|
}, |
|
|
) |
|
|
self._commit_snapshot_id = snapshot_id |
|
|
self._sequence_number = sequence_number |
|
|
|
|
|
def prepare_manifest(self, manifest_file: ManifestFile) -> ManifestFile: |
|
|
wrapped_manifest_file = copy(manifest_file) |
|
|
|
|
|
if wrapped_manifest_file.sequence_number == UNASSIGNED_SEQ: |
|
|
|
|
|
|
|
|
if self._commit_snapshot_id != wrapped_manifest_file.added_snapshot_id: |
|
|
raise ValueError( |
|
|
f"Found unassigned sequence number for a manifest from snapshot: {self._commit_snapshot_id} != {wrapped_manifest_file.added_snapshot_id}" |
|
|
) |
|
|
wrapped_manifest_file.sequence_number = self._sequence_number |
|
|
|
|
|
if wrapped_manifest_file.min_sequence_number == UNASSIGNED_SEQ: |
|
|
if self._commit_snapshot_id != wrapped_manifest_file.added_snapshot_id: |
|
|
raise ValueError( |
|
|
f"Found unassigned sequence number for a manifest from snapshot: {wrapped_manifest_file.added_snapshot_id}" |
|
|
) |
|
|
|
|
|
|
|
|
wrapped_manifest_file.min_sequence_number = self._sequence_number |
|
|
return wrapped_manifest_file |
|
|
|
|
|
|
|
|
def write_manifest_list( |
|
|
format_version: TableVersion, |
|
|
output_file: OutputFile, |
|
|
snapshot_id: int, |
|
|
parent_snapshot_id: Optional[int], |
|
|
sequence_number: Optional[int], |
|
|
) -> ManifestListWriter: |
|
|
if format_version == 1: |
|
|
return ManifestListWriterV1(output_file, snapshot_id, parent_snapshot_id) |
|
|
elif format_version == 2: |
|
|
if sequence_number is None: |
|
|
raise ValueError(f"Sequence-number is required for V2 tables: {sequence_number}") |
|
|
return ManifestListWriterV2(output_file, snapshot_id, parent_snapshot_id, sequence_number) |
|
|
else: |
|
|
raise ValueError(f"Cannot write manifest list for table version: {format_version}") |
|
|
|