|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import sys |
|
|
from collections.abc import Mapping |
|
|
from typing import TYPE_CHECKING, Optional |
|
|
|
|
|
import numpy as np |
|
|
import pyarrow as pa |
|
|
|
|
|
from .. import config |
|
|
from ..utils.logging import get_logger |
|
|
from ..utils.py_utils import map_nested |
|
|
from .formatting import TensorFormatter |
|
|
|
|
|
|
|
|
if TYPE_CHECKING: |
|
|
import jax |
|
|
import jaxlib |
|
|
|
|
|
logger = get_logger() |
|
|
|
|
|
DEVICE_MAPPING: Optional[dict] = None |
|
|
|
|
|
|
|
|
class JaxFormatter(TensorFormatter[Mapping, "jax.Array", Mapping]): |
|
|
def __init__(self, features=None, device=None, token_per_repo_id=None, **jnp_array_kwargs): |
|
|
super().__init__(features=features, token_per_repo_id=token_per_repo_id) |
|
|
import jax |
|
|
from jaxlib.xla_client import Device |
|
|
|
|
|
if isinstance(device, Device): |
|
|
raise ValueError( |
|
|
f"Expected {device} to be a `str` not {type(device)}, as `jaxlib.xla_extension.Device` " |
|
|
"is not serializable neither with `pickle` nor with `dill`. Instead you can surround " |
|
|
"the device with `str()` to get its string identifier that will be internally mapped " |
|
|
"to the actual `jaxlib.xla_extension.Device`." |
|
|
) |
|
|
self.device = device if isinstance(device, str) else str(jax.devices()[0]) |
|
|
|
|
|
|
|
|
global DEVICE_MAPPING |
|
|
if DEVICE_MAPPING is None: |
|
|
DEVICE_MAPPING = self._map_devices_to_str() |
|
|
if self.device not in list(DEVICE_MAPPING.keys()): |
|
|
logger.warning( |
|
|
f"Device with string identifier {self.device} not listed among the available " |
|
|
f"devices: {list(DEVICE_MAPPING.keys())}, so falling back to the default " |
|
|
f"device: {str(jax.devices()[0])}." |
|
|
) |
|
|
self.device = str(jax.devices()[0]) |
|
|
self.jnp_array_kwargs = jnp_array_kwargs |
|
|
|
|
|
@staticmethod |
|
|
def _map_devices_to_str() -> dict[str, "jaxlib.xla_extension.Device"]: |
|
|
import jax |
|
|
|
|
|
return {str(device): device for device in jax.devices()} |
|
|
|
|
|
def _consolidate(self, column): |
|
|
import jax |
|
|
import jax.numpy as jnp |
|
|
|
|
|
if isinstance(column, list) and column: |
|
|
if all( |
|
|
isinstance(x, jax.Array) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column |
|
|
): |
|
|
return jnp.stack(column, axis=0) |
|
|
return column |
|
|
|
|
|
def _tensorize(self, value): |
|
|
import jax |
|
|
import jax.numpy as jnp |
|
|
|
|
|
if isinstance(value, (str, bytes, type(None))): |
|
|
return value |
|
|
elif isinstance(value, (np.character, np.ndarray)) and np.issubdtype(value.dtype, np.character): |
|
|
return value.tolist() |
|
|
|
|
|
default_dtype = {} |
|
|
|
|
|
if isinstance(value, (np.number, np.ndarray)) and np.issubdtype(value.dtype, np.integer): |
|
|
|
|
|
|
|
|
if jax.config.jax_enable_x64: |
|
|
default_dtype = {"dtype": jnp.int64} |
|
|
else: |
|
|
default_dtype = {"dtype": jnp.int32} |
|
|
elif isinstance(value, (np.number, np.ndarray)) and np.issubdtype(value.dtype, np.floating): |
|
|
default_dtype = {"dtype": jnp.float32} |
|
|
|
|
|
if config.PIL_AVAILABLE and "PIL" in sys.modules: |
|
|
import PIL.Image |
|
|
|
|
|
if isinstance(value, PIL.Image.Image): |
|
|
value = np.asarray(value) |
|
|
if config.TORCHVISION_AVAILABLE and "torchvision" in sys.modules: |
|
|
from torchvision.io import VideoReader |
|
|
|
|
|
if isinstance(value, VideoReader): |
|
|
return value |
|
|
if config.TORCHCODEC_AVAILABLE and "torchcodec" in sys.modules: |
|
|
from torchcodec.decoders import AudioDecoder, VideoDecoder |
|
|
|
|
|
if isinstance(value, (VideoDecoder, AudioDecoder)): |
|
|
return value |
|
|
|
|
|
|
|
|
|
|
|
global DEVICE_MAPPING |
|
|
if DEVICE_MAPPING is None: |
|
|
DEVICE_MAPPING = self._map_devices_to_str() |
|
|
|
|
|
with jax.default_device(DEVICE_MAPPING[self.device]): |
|
|
|
|
|
|
|
|
return jnp.array(value, **{**default_dtype, **self.jnp_array_kwargs}) |
|
|
|
|
|
def _recursive_tensorize(self, data_struct): |
|
|
import jax |
|
|
|
|
|
|
|
|
if config.TORCH_AVAILABLE and "torch" in sys.modules: |
|
|
import torch |
|
|
|
|
|
if isinstance(data_struct, torch.Tensor): |
|
|
return self._tensorize(data_struct.detach().cpu().numpy()[()]) |
|
|
if hasattr(data_struct, "__array__") and not isinstance(data_struct, jax.Array): |
|
|
data_struct = data_struct.__array__() |
|
|
|
|
|
if isinstance(data_struct, np.ndarray): |
|
|
if data_struct.dtype == object: |
|
|
return self._consolidate([self.recursive_tensorize(substruct) for substruct in data_struct]) |
|
|
elif isinstance(data_struct, (list, tuple)): |
|
|
return self._consolidate([self.recursive_tensorize(substruct) for substruct in data_struct]) |
|
|
return self._tensorize(data_struct) |
|
|
|
|
|
def recursive_tensorize(self, data_struct: dict): |
|
|
return map_nested(self._recursive_tensorize, data_struct, map_list=False) |
|
|
|
|
|
def format_row(self, pa_table: pa.Table) -> Mapping: |
|
|
row = self.numpy_arrow_extractor().extract_row(pa_table) |
|
|
row = self.python_features_decoder.decode_row(row) |
|
|
return self.recursive_tensorize(row) |
|
|
|
|
|
def format_column(self, pa_table: pa.Table) -> "jax.Array": |
|
|
column = self.numpy_arrow_extractor().extract_column(pa_table) |
|
|
column = self.python_features_decoder.decode_column(column, pa_table.column_names[0]) |
|
|
column = self.recursive_tensorize(column) |
|
|
column = self._consolidate(column) |
|
|
return column |
|
|
|
|
|
def format_batch(self, pa_table: pa.Table) -> Mapping: |
|
|
batch = self.numpy_arrow_extractor().extract_batch(pa_table) |
|
|
batch = self.python_features_decoder.decode_batch(batch) |
|
|
batch = self.recursive_tensorize(batch) |
|
|
for column_name in batch: |
|
|
batch[column_name] = self._consolidate(batch[column_name]) |
|
|
return batch |
|
|
|