hc99's picture
Add files using upload-large-folder tool
2c3c408 verified
# ruff: noqa: F811
import base64
import gc
import io
import os
import pickle
import random
import sys
import tarfile
import time
import warnings
from collections import OrderedDict
import numpy as np
import psutil
import pytest
from numpy.testing import assert_array_equal
import tiledb
from tiledb.datatypes import DataType
from .common import (
DiskTestCase,
assert_captured,
assert_subarrays_equal,
assert_unordered_equal,
fx_sparse_cell_order, # noqa: F401
has_pandas,
has_pyarrow,
rand_ascii,
rand_ascii_bytes,
rand_utf8,
)
@pytest.fixture(scope="class")
def test_incomplete_return_array(tmpdir_factory, request):
tmp_path = str(tmpdir_factory.mktemp("array"))
ncells = 20
nvals = 10
data = np.array([rand_utf8(nvals - i % 2) for i in range(ncells)], dtype="O")
dom = tiledb.Domain(tiledb.Dim(domain=(0, len(data) - 1), tile=len(data)))
att = tiledb.Attr(dtype=str, var=True)
allows_duplicates = request.param
schema = tiledb.ArraySchema(
dom, (att,), sparse=True, allows_duplicates=allows_duplicates
)
coords = np.arange(ncells)
tiledb.SparseArray.create(tmp_path, schema)
with tiledb.SparseArray(tmp_path, mode="w") as T:
T[coords] = data
with tiledb.SparseArray(tmp_path, mode="r") as T:
assert_subarrays_equal(data, T[:][""])
return tmp_path
class VersionTest(DiskTestCase):
def test_libtiledb_version(self):
v = tiledb.libtiledb.version()
self.assertIsInstance(v, tuple)
self.assertTrue(len(v) == 3)
self.assertTrue(v[0] >= 1, "TileDB major version must be >= 1")
def test_tiledbpy_version(self):
v = tiledb.version.version
self.assertIsInstance(v, str)
v = tiledb.version()
self.assertIsInstance(v, tuple)
self.assertTrue(3 <= len(v) <= 5)
class ArrayTest(DiskTestCase):
def create_array_schema(self):
filters = tiledb.FilterList([tiledb.ZstdFilter(level=5)])
domain = tiledb.Domain(
tiledb.Dim(domain=(1, 8), tile=2, filters=filters),
tiledb.Dim(domain=(1, 8), tile=2, filters=filters),
)
a1 = tiledb.Attr("val", dtype="f8")
return tiledb.ArraySchema(domain=domain, attrs=(a1,))
def test_array_create(self):
config = tiledb.Config()
config["sm.consolidation.step_min_frag"] = 0
config["sm.consolidation.steps"] = 1
schema = self.create_array_schema()
# persist array schema
tiledb.libtiledb.Array.create(self.path("foo"), schema)
# these should be no-ops
# full signature
tiledb.consolidate(self.path("foo"), config=config)
# kw signature
tiledb.consolidate(uri=self.path("foo"))
# load array in readonly mode
array = tiledb.libtiledb.Array(self.path("foo"), mode="r")
self.assertTrue(array.isopen)
self.assertEqual(array.schema, schema)
self.assertEqual(array.mode, "r")
self.assertEqual(array.uri, self.path("foo"))
# test that we cannot consolidate an array in readonly mode
with self.assertRaises(tiledb.TileDBError):
array.consolidate(config=config)
# we have not written anything, so the array is empty
self.assertIsNone(array.nonempty_domain())
array.reopen()
self.assertTrue(array.isopen)
array.close()
self.assertEqual(array.isopen, False)
with self.assertRaises(tiledb.TileDBError):
# cannot get schema from closed array
array.schema
with self.assertRaises(tiledb.TileDBError):
# cannot re-open a closed array
array.reopen()
def test_array_create_with_ctx(self):
schema = self.create_array_schema()
with self.assertRaises(TypeError):
tiledb.libtiledb.Array.create(self.path("foo"), schema, ctx="foo")
# persist array schema
tiledb.libtiledb.Array.create(self.path("foo"), schema, ctx=tiledb.Ctx())
@pytest.mark.skipif(
not (sys.platform == "win32" and tiledb.libtiledb.version() >= (2, 3, 0)),
reason="Shared network drive only on Win32",
)
def test_array_create_on_shared_drive(self):
schema = self.create_array_schema()
uri = self.path(basename="foo", shared=True)
tiledb.libtiledb.Array.create(uri, schema)
# load array in readonly mode
array = tiledb.libtiledb.Array(uri, mode="r")
self.assertTrue(array.isopen)
self.assertEqual(array.schema, schema)
self.assertEqual(array.mode, "r")
self.assertEqual(array.uri, uri)
# we have not written anything, so the array is empty
self.assertIsNone(array.nonempty_domain())
array.reopen()
self.assertTrue(array.isopen)
array.close()
self.assertEqual(array.isopen, False)
with self.assertRaises(tiledb.TileDBError):
# cannot get schema from closed array
array.schema
with self.assertRaises(tiledb.TileDBError):
# cannot re-open a closed array
array.reopen()
def test_array_create_encrypted(self):
config = tiledb.Config()
config["sm.consolidation.step_min_frags"] = 0
config["sm.consolidation.steps"] = 1
schema = self.create_array_schema()
key = "0123456789abcdeF0123456789abcdeF"
# persist array schema
tiledb.libtiledb.Array.create(self.path("foo"), schema, key=key)
# check that we can open the array sucessfully
config = tiledb.Config()
config["sm.encryption_key"] = key
config["sm.encryption_type"] = "AES_256_GCM"
ctx = tiledb.Ctx(config=config)
with tiledb.libtiledb.Array(self.path("foo"), mode="r", ctx=ctx) as array:
self.assertTrue(array.isopen)
self.assertEqual(array.schema, schema)
self.assertEqual(array.mode, "r")
with tiledb.open(self.path("foo"), mode="r", key=key, ctx=ctx) as array:
self.assertTrue(array.isopen)
self.assertEqual(array.schema, schema)
self.assertEqual(array.mode, "r")
tiledb.consolidate(uri=self.path("foo"), ctx=tiledb.Ctx(config))
config = tiledb.Config()
config["sm.encryption_key"] = "0123456789abcdeF0123456789abcdeX"
config["sm.encryption_type"] = "AES_256_GCM"
ctx = tiledb.Ctx(config=config)
# check that opening the array with the wrong key fails:
with self.assertRaises(tiledb.TileDBError):
tiledb.libtiledb.Array(self.path("foo"), mode="r", ctx=ctx)
config = tiledb.Config()
config["sm.encryption_key"] = "0123456789abcdeF0123456789abcde"
config["sm.encryption_type"] = "AES_256_GCM"
ctx = tiledb.Ctx(config=config)
# check that opening the array with the wrong key length fails:
with self.assertRaises(tiledb.TileDBError):
tiledb.libtiledb.Array(self.path("foo"), mode="r", ctx=ctx)
config = tiledb.Config()
config["sm.encryption_key"] = "0123456789abcdeF0123456789abcde"
config["sm.encryption_type"] = "AES_256_GCM"
ctx = tiledb.Ctx(config=config)
# check that consolidating the array with the wrong key fails:
with self.assertRaises(tiledb.TileDBError):
tiledb.consolidate(self.path("foo"), config=config, ctx=ctx)
# needs core fix in 2.2.4
@pytest.mark.skipif(
(sys.platform == "win32" and tiledb.libtiledb.version() == (2, 2, 3)),
reason="Skip array_doesnt_exist test on Win32 / libtiledb 2.2.3",
)
def test_array_doesnt_exist(self):
with self.assertRaises(tiledb.TileDBError):
tiledb.libtiledb.Array(self.path("foo"), mode="r")
def test_create_schema_matches(self):
dims = (tiledb.Dim(domain=(0, 6), tile=2),)
dom = tiledb.Domain(*dims)
att = tiledb.Attr(dtype=np.byte)
schema = tiledb.ArraySchema(domain=dom, attrs=(att,), sparse=True)
uri = self.path("s1")
with self.assertRaises(ValueError):
tiledb.DenseArray.create(uri, schema)
dense_schema = tiledb.ArraySchema(domain=dom, attrs=(att,))
uri = self.path("d1")
with self.assertRaises(ValueError):
tiledb.SparseArray.create(uri, dense_schema)
class MySparseArray(tiledb.SparseArray):
pass
with self.assertRaises(ValueError):
MySparseArray.create(uri, dense_schema)
def test_nonempty_domain_scalar(self):
uri = self.path("test_nonempty_domain_scalar")
dims = tiledb.Dim(domain=(-10, 10), dtype=np.int64, tile=1)
schema = tiledb.ArraySchema(
tiledb.Domain(dims), attrs=[tiledb.Attr(dtype=np.int32)], sparse=True
)
tiledb.Array.create(uri, schema)
with tiledb.open(uri, "w") as A:
A[-1] = 10
A[1] = 11
with tiledb.open(uri, "r") as A:
ned = A.nonempty_domain()
assert_array_equal(ned, ((-1, 1),))
assert isinstance(ned[0][0], int)
assert isinstance(ned[0][1], int)
def test_nonempty_domain_empty_string(self):
uri = self.path("test_nonempty_domain_empty_string")
dims = [tiledb.Dim(dtype="ascii")]
schema = tiledb.ArraySchema(tiledb.Domain(dims), sparse=True)
tiledb.Array.create(uri, schema)
with tiledb.open(uri, "r") as A:
assert_array_equal(A.nonempty_domain(), ((None, None),))
with tiledb.open(uri, "w") as A:
A[""] = None
with tiledb.open(uri, "r") as A:
assert_array_equal(A.nonempty_domain(), ((b"", b""),))
def test_create_array_overwrite(self):
uri = self.path("test_create_array_overwrite")
dims = tiledb.Dim(domain=(0, 10), dtype=np.int64)
schema = tiledb.ArraySchema(
tiledb.Domain(dims), attrs=[tiledb.Attr(dtype=np.int32)], sparse=True
)
with pytest.warns(UserWarning, match="Overwrite set, but array does not exist"):
tiledb.Array.create(uri, schema, overwrite=True)
with tiledb.open(uri, "w") as A:
A[0] = 1
with tiledb.open(uri, "r") as A:
assert A.nonempty_domain() == ((0, 0),)
# cannot overwrite the array by default
with self.assertRaises(tiledb.TileDBError):
tiledb.Array.create(uri, schema)
tiledb.Array.create(uri, schema, overwrite=True)
# make the old array has been deleted and replaced
with tiledb.open(uri, "r") as A:
assert A.nonempty_domain() is None
def test_upgrade_version(self):
tgz_sparse = b"""H4sIAJKNpWAAA+2aPW/TQBjHz2nTFlGJClUoAxIuA0ICpXf2vdhbGWBiYEIgihI7MRT1JVKairKh
qgNfgA2kDnwFVga+ABtfgE8AEwsS5/ROTUzBjWpbKv3/JPexLxc/l/59zz3PJc1lUjqUUiWEO7Ty
0GqsPbxgnArmUymk71LmUc6JK8ofGiE724Oor4fyYqvXH/S2/trv5VqSbPzjPuMfyi18nCXRXG61
NpNBVOZjMIH+XEip9fc9xaB/FaT6b/Q6681BNy7Lh/5/SM4n0l8JPf9pWQMaBfq3on4/etXa7qwl
m1EZz0Ge/p6X1V9wKaF/FdT1sWrOXxs77dhXLw//OiRtcNKuzvBspH+gjwVz7Yy07TqdhNTuzcw4
OwtT0407qzM3Hi58vzZH7678cN99rl9f2ji40JZ77T0Wzb+JD/rdp8SZnfta2gcFx5LOfyY9xqXn
ByoIVeYqDJMu44GOyGHCeRIGKuHCF1HsRRGLaacl8jOHifM/z2M+8r9KOL3+zd56jo8J1n+rPxcC
8b8KjvRnvlSh8rJXcRJ2Euor7gne8XgsJdVPhAoSFXZFogrWX6//aqg/p9C/Ck6vf6Hx3+rPmEL8
r4IC9G+1nvWj55vJ1mC4k9CNBpkqImf+a7VFRn8phI/5XwVpUh+Yc9fYk+b/af9FMp7/27Zd51vc
brf3Y7c+e//BFeJ8IJfSG9hoYd9zUl9p/4sZX7ZN1xrdlXrquwYXcAEXx7s4ojbSOWXK2NtknBVy
Mmxc/GKsZ2781tifxj4xjj8Zu2Qc79sBgKopYP3v5u0Z5uX/7I/8z6ce9n8rwYaAhj6ukvE4Yttu
flz+5TbeE/JIt9vYUSO3Hs8Pwww4wxQw/3O/Msit/wXP1n9Sof6vhNH538i02ak+njyA/4kC9v+L
rP/N/q8UmP/VgPofLuDiXLg4AvU/MBSw/hdZ/5v1XxcCDOt/FaD+P98UMP+LrP/t7z8Uxe8/KgH1
PwAAAAAAAAAAAAAAAAAAAAAAAHD2+Q18oX51AFAAAA=="""
path = self.path("test_upgrade_version")
with tarfile.open(fileobj=io.BytesIO(base64.b64decode(tgz_sparse))) as tf:
try:
tf.extractall(path, filter="fully_trusted")
except TypeError:
tf.extractall(path)
with tiledb.open(path) as A:
assert A.schema.version == 5
A.upgrade_version()
with tiledb.open(path) as A:
assert A.schema.version >= 15
@pytest.mark.parametrize("use_timestamps", [True, False])
def test_array_delete_fragments(self, use_timestamps):
dshape = (1, 3)
num_writes = 10
def create_array(target_path, dshape):
dom = tiledb.Domain(tiledb.Dim(domain=dshape, tile=len(dshape)))
att = tiledb.Attr(dtype="int64")
schema = tiledb.ArraySchema(domain=dom, attrs=(att,), sparse=True)
tiledb.libtiledb.Array.create(target_path, schema)
def write_fragments(target_path, dshape, num_writes):
for i in range(1, num_writes + 1):
with tiledb.open(
target_path, "w", timestamp=i if use_timestamps else None
) as A:
A[[1, 2, 3]] = np.random.rand(dshape[1])
path = self.path("test_array_delete_fragments")
ts = tuple((t, t) for t in range(1, 11))
create_array(path, dshape)
write_fragments(path, dshape, num_writes)
frags = tiledb.array_fragments(path)
assert len(frags) == 10
if use_timestamps:
assert frags.timestamp_range == ts
if use_timestamps:
tiledb.Array.delete_fragments(path, 3, 6)
else:
timestamps = [t[0] for t in tiledb.array_fragments(path).timestamp_range]
tiledb.Array.delete_fragments(path, timestamps[2], timestamps[5])
frags = tiledb.array_fragments(path)
assert len(frags) == 6
if use_timestamps:
assert frags.timestamp_range == ts[:2] + ts[6:]
def test_array_delete(self):
uri = self.path("test_array_delete")
data = np.random.rand(10)
tiledb.from_numpy(uri, data)
with tiledb.open(uri) as A:
assert_array_equal(A[:], data)
assert tiledb.array_exists(uri) is True
tiledb.Array.delete_array(uri)
assert tiledb.array_exists(uri) is False
@pytest.mark.skipif(
not has_pyarrow() or not has_pandas(),
reason="pyarrow>=1.0 and/or pandas>=1.0,<3.0 not installed",
)
@pytest.mark.parametrize("sparse", [True, False])
@pytest.mark.parametrize("pass_df", [True, False])
def test_array_write_nullable(self, sparse, pass_df):
import pyarrow as pa
uri = self.path("test_array_write_nullable")
dom = tiledb.Domain(tiledb.Dim("d", domain=(1, 5), dtype="int64"))
att1 = tiledb.Attr("a1", dtype="int8", nullable=True)
att2 = tiledb.Attr("a2", dtype="str", nullable=True)
schema = tiledb.ArraySchema(domain=dom, attrs=[att1, att2], sparse=sparse)
tiledb.Array.create(uri, schema)
with tiledb.open(uri, "w") as A:
dims = pa.array([1, 2, 3, 4, 5])
data1 = pa.array([1.0, 2.0, None, 0, 1.0])
data2 = pa.array(["a", "b", None, None, "c"])
if pass_df:
dims = dims.to_pandas()
data1 = data1.to_pandas()
data2 = data2.to_pandas()
if sparse:
A[dims] = {"a1": data1, "a2": data2}
else:
A[:] = {"a1": data1, "a2": data2}
with tiledb.open(uri, "r") as A:
expected_validity1 = [False, False, True, False, False]
assert_array_equal(A[:]["a1"].mask, expected_validity1)
assert_array_equal(A.df[:]["a1"].isna(), expected_validity1)
assert_array_equal(A.query(attrs=["a1"])[:]["a1"].mask, expected_validity1)
expected_validity2 = [False, False, True, True, False]
assert_array_equal(A[:]["a2"].mask, expected_validity2)
assert_array_equal(A.df[:]["a2"].isna(), expected_validity2)
assert_array_equal(A.query(attrs=["a2"])[:]["a2"].mask, expected_validity2)
with tiledb.open(uri, "w") as A:
dims = pa.array([1, 2, 3, 4, 5])
data1 = pa.array([None, None, None, None, None])
data2 = pa.array([None, None, None, None, None])
if pass_df:
dims = dims.to_pandas()
data1 = data1.to_pandas()
data2 = data2.to_pandas()
if sparse:
A[dims] = {"a1": data1, "a2": data2}
else:
A[:] = {"a1": data1, "a2": data2}
with tiledb.open(uri, "r") as A:
expected_validity1 = [True, True, True, True, True]
assert_array_equal(A[:]["a1"].mask, expected_validity1)
assert_array_equal(A.df[:]["a1"].isna(), expected_validity1)
assert_array_equal(A.query(attrs=["a1"])[:]["a1"].mask, expected_validity1)
expected_validity2 = [True, True, True, True, True]
assert_array_equal(A[:]["a2"].mask, expected_validity2)
assert_array_equal(A.df[:]["a2"].isna(), expected_validity2)
assert_array_equal(A.query(attrs=["a2"])[:]["a2"].mask, expected_validity2)
class DenseArrayTest(DiskTestCase):
def test_array_1d(self):
A = np.arange(1050)
dom = tiledb.Domain(tiledb.Dim(domain=(0, 1049), tile=100, dtype=np.int64))
att = tiledb.Attr(dtype=A.dtype)
schema = tiledb.ArraySchema(domain=dom, attrs=(att,))
tiledb.DenseArray.create(self.path("foo"), schema)
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
self.assertEqual(len(A), len(T))
self.assertEqual(A.ndim, T.ndim)
self.assertEqual(A.shape, T.shape)
self.assertEqual(1, T.nattr)
self.assertEqual(A.dtype, T.attr(0).dtype)
self.assertEqual(T.dim(T.schema.domain.dim(0).name), T.dim(0))
with self.assertRaises(ValueError):
T.dim(1.0)
self.assertIsInstance(T.timestamp_range, tuple)
self.assertTrue(T.timestamp_range[1] > 0)
# check empty array
B = T[:]
self.assertEqual(A.shape, B.shape)
self.assertEqual(A.dtype, B.dtype)
self.assertIsNone(T.nonempty_domain())
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
# check set array
T[:] = A
read1_timestamp = -1
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
self.assertEqual(((0, 1049),), T.nonempty_domain())
# check timestamp
read1_timestamp = T.timestamp_range
self.assertTrue(read1_timestamp[1] > 0)
# check slicing
assert_array_equal(A, np.array(T))
assert_array_equal(A, T[:])
assert_array_equal(A, T[...])
assert_array_equal(A, T[slice(None)])
assert_array_equal(A[:10], T[:10])
assert_array_equal(A[10:20], T[10:20])
assert_array_equal(A[-10:], T[-10:])
# ellipsis
assert_array_equal(A[:10, ...], T[:10, ...])
assert_array_equal(A[10:50, ...], T[10:50, ...])
assert_array_equal(A[-50:, ...], T[-50:, ...])
assert_array_equal(A[..., :10], T[..., :10])
assert_array_equal(A[..., 10:20], T[..., 10:20])
assert_array_equal(A[..., -50:], T[..., -50:])
# across tiles
assert_array_equal(A[:150], T[:150])
assert_array_equal(A[-250:], T[-250:])
# point index
self.assertEqual(A[0], T[0])
self.assertEqual(A[-1], T[-1])
# point index with all index types
self.assertEqual(A[123], T[np.int8(123)])
self.assertEqual(A[123], T[np.uint8(123)])
self.assertEqual(A[123], T[np.int16(123)])
self.assertEqual(A[123], T[np.uint16(123)])
self.assertEqual(A[123], T[np.int64(123)])
self.assertEqual(A[123], T[np.uint64(123)])
self.assertEqual(A[123], T[np.int32(123)])
self.assertEqual(A[123], T[np.uint32(123)])
# mixed-type slicing
# https://github.com/TileDB-Inc/TileDB-Py/issues/140
self.assertEqual(A[0:1], T[0 : np.uint16(1)])
self.assertEqual(A[0:1], T[np.int64(0) : 1])
with self.assertRaises(IndexError):
# this is a consequence of NumPy promotion rules
self.assertEqual(A[0:1], T[np.uint64(0) : 1])
# basic step
assert_array_equal(A[:50:2], T[:50:2])
assert_array_equal(A[:2:50], T[:2:50])
assert_array_equal(A[10:-1:50], T[10:-1:50])
# indexing errors
with self.assertRaises(IndexError):
T[:, :]
with self.assertRaises(IndexError):
T[:, 50]
with self.assertRaises(IndexError):
T[50, :]
with self.assertRaises(IndexError):
T[0, 0]
# check single ellipsis
with self.assertRaises(IndexError):
T[..., 1:5, ...]
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
# check partial assignment
B = np.arange(1e5, 2e5).astype(A.dtype)
T[190:310] = B[190:310]
read2_timestamp = -1
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
assert_array_equal(A[:190], T[:190])
assert_array_equal(B[190:310], T[190:310])
assert_array_equal(A[310:], T[310:])
# test timestamps are updated
read2_timestamp = T.timestamp_range
self.assertTrue(read2_timestamp > read1_timestamp)
def test_array_1d_set_scalar(self):
A = np.zeros(50)
dom = tiledb.Domain(tiledb.Dim(domain=(0, 49), tile=50))
att = tiledb.Attr(dtype=A.dtype)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.DenseArray.create(self.path("foo"), schema)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = A
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
assert_array_equal(A, T[:])
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
value = -1, 3, 10
A[0], A[1], A[3] = value
T[0], T[1], T[3] = value
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
assert_array_equal(A, T[:])
for value in (-1, 3, 10):
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
A[5:25] = value
T[5:25] = value
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
assert_array_equal(A, T[:])
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
A[:] = value
T[:] = value
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
assert_array_equal(A, T[:])
def test_array_id_point_queries(self):
# TODO: handle queries like T[[2, 5, 10]] = ?
pass
@pytest.mark.parametrize("dtype", ["u1", "u2", "u4", "u8", "i1", "i2", "i4", "i8"])
def test_dense_index_dtypes(self, dtype):
path = self.path()
data = np.arange(0, 3).astype(dtype)
with tiledb.from_numpy(path, data):
pass
with tiledb.open(path) as B:
assert_array_equal(B[:], data)
@pytest.mark.skipif(
tiledb.libtiledb.version() < (2, 10),
reason="TILEDB_BOOL introduced in libtiledb 2.10",
)
def test_dense_index_bool(self):
path = self.path()
data = np.random.randint(0, 1, 10, dtype=bool)
with tiledb.from_numpy(path, data):
pass
with tiledb.open(path) as B:
assert_array_equal(B[:], data)
def test_array_2d(self):
A = np.arange(10000).reshape((1000, 10))
dom = tiledb.Domain(
tiledb.Dim(domain=(0, 999), tile=100), tiledb.Dim(domain=(0, 9), tile=2)
)
att = tiledb.Attr(dtype=A.dtype)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.DenseArray.create(self.path("foo"), schema)
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
self.assertEqual(len(A), len(T))
self.assertEqual(A.ndim, T.ndim)
self.assertEqual(A.shape, T.shape)
self.assertEqual(1, T.nattr)
self.assertEqual(A.dtype, T.attr(0).dtype)
# check that the non-empty domain is None
self.assertIsNone(T.nonempty_domain())
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
# Set data
T[:] = A
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
assert_array_equal(A, T[:])
# check the non-empty domain spans the whole domain
self.assertEqual(((0, 999), (0, 9)), T.nonempty_domain())
# check array-like
assert_array_equal(A, np.array(T))
# slicing
assert_array_equal(A, T[:])
assert_array_equal(A, T[...])
assert_array_equal(A, T[slice(None)])
# slice first dimension
assert_array_equal(A[:10], T[:10])
assert_array_equal(A[:10], T[:10])
assert_array_equal(A[10:20], T[10:20])
assert_array_equal(A[-10:], T[-10:])
assert_array_equal(A[:10, :], T[:10, :])
assert_array_equal(A[10:20, :], T[10:20, :])
assert_array_equal(A[-10:, :], T[-10:, :])
assert_array_equal(A[:10, ...], T[:10, ...])
assert_array_equal(A[10:20, ...], T[10:20, ...])
assert_array_equal(A[-10:, ...], T[-10:, ...])
assert_array_equal(A[:10, :, ...], T[:10, :, ...])
assert_array_equal(A[10:20, :, ...], T[10:20, :, ...])
assert_array_equal(A[-10:, :, ...], T[-10:, :, ...])
# slice second dimension
assert_array_equal(A[:, :2], T[:, :2])
assert_array_equal(A[:, 2:4], T[:, 2:4])
assert_array_equal(A[:, -2:], T[:, -2:])
assert_array_equal(A[..., :2], T[..., :2])
assert_array_equal(A[..., 2:4], T[..., 2:4])
assert_array_equal(A[..., -2:], T[..., -2:])
assert_array_equal(A[:, ..., :2], T[:, ..., :2])
assert_array_equal(A[:, ..., 2:4], T[:, ..., 2:4])
assert_array_equal(A[:, ..., -2:], T[:, ..., -2:])
# slice both dimensions
assert_array_equal(A[:10, :2], T[:10, :2])
assert_array_equal(A[10:20, 2:4], T[10:20, 2:4])
assert_array_equal(A[-10:, -2:], T[-10:, -2:])
# slice across tile boundries
assert_array_equal(A[:110], T[:110])
assert_array_equal(A[190:310], T[190:310])
assert_array_equal(A[-110:], T[-110:])
assert_array_equal(A[:110, :], T[:110, :])
assert_array_equal(A[190:310, :], T[190:310, :])
assert_array_equal(A[-110:, :], T[-110:, :])
assert_array_equal(A[:, :3], T[:, :3])
assert_array_equal(A[:, 3:7], T[:, 3:7])
assert_array_equal(A[:, -3:], T[:, -3:])
assert_array_equal(A[:110, :3], T[:110, :3])
assert_array_equal(A[190:310, 3:7], T[190:310, 3:7])
assert_array_equal(A[-110:, -3:], T[-110:, -3:])
# single row/col/item
assert_array_equal(A[0], T[0])
assert_array_equal(A[-1], T[-1])
assert_array_equal(A[:, 0], T[:, 0])
assert_array_equal(A[:, -1], T[:, -1])
self.assertEqual(A[0, 0], T[0, 0])
self.assertEqual(A[-1, -1], T[-1, -1])
# too many indices
with self.assertRaises(IndexError):
T[:, :, :]
with self.assertRaises(IndexError):
T[0, :, :]
with self.assertRaises(IndexError):
T[:, 0, :]
with self.assertRaises(IndexError):
T[:, :, 0]
with self.assertRaises(IndexError):
T[0, 0, 0]
# only single ellipsis allowed
with self.assertRaises(IndexError):
T[..., ...]
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
# check partial assignment
B = np.arange(10000, 20000).reshape((1000, 10))
T[190:310, 3:7] = B[190:310, 3:7]
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
assert_array_equal(A[:190], T[:190])
assert_array_equal(A[:, :3], T[:, :3])
assert_array_equal(B[190:310, 3:7], T[190:310, 3:7])
assert_array_equal(A[310:], T[310:])
assert_array_equal(A[:, 7:], T[:, 7:])
@pytest.mark.skipif(
not (sys.platform == "win32" and tiledb.libtiledb.version() >= (2, 3, 0)),
reason="Shared network drive only on Win32",
)
def test_array_1d_shared_drive(self):
A = np.zeros(50)
dom = tiledb.Domain(tiledb.Dim(domain=(0, 49), tile=50))
att = tiledb.Attr(dtype=A.dtype)
schema = tiledb.ArraySchema(dom, (att,))
uri = self.path("foo", shared=True)
tiledb.DenseArray.create(uri, schema)
with tiledb.DenseArray(uri, mode="w") as T:
T[:] = A
with tiledb.DenseArray(uri, mode="r") as T:
assert_array_equal(A, T[:])
with tiledb.DenseArray(uri, mode="w") as T:
value = -1, 3, 10
A[0], A[1], A[3] = value
T[0], T[1], T[3] = value
with tiledb.DenseArray(uri, mode="r") as T:
assert_array_equal(A, T[:])
for value in (-1, 3, 10):
with tiledb.DenseArray(uri, mode="w") as T:
A[5:25] = value
T[5:25] = value
with tiledb.DenseArray(uri, mode="r") as T:
assert_array_equal(A, T[:])
with tiledb.DenseArray(uri, mode="w") as T:
A[:] = value
T[:] = value
with tiledb.DenseArray(uri, mode="r") as T:
assert_array_equal(A, T[:])
def test_fixed_string(self):
a = np.array(["ab", "cd", "ef", "gh", "ij", "kl", "", "op"], dtype="|S2")
with tiledb.from_numpy(self.path("fixed_string"), a) as T:
with tiledb.open(self.path("fixed_string")) as R:
self.assertEqual(T.dtype, R.dtype)
self.assertEqual(R.attr(0).ncells, 2)
assert_array_equal(T, R)
def test_ncell_int(self):
a = np.array([(1, 2), (3, 4), (5, 6)], dtype=[("", np.int16), ("", np.int16)])
with tiledb.from_numpy(self.path("ncell_int16"), a) as T:
with tiledb.open(self.path("ncell_int16")) as R:
self.assertEqual(T.dtype, R.dtype)
self.assertEqual(R.attr(0).ncells, 2)
assert_array_equal(T, R)
assert_array_equal(T, R.multi_index[0:2][""])
@pytest.mark.parametrize("use_timestamps", [True, False])
def test_open_with_timestamp(self, use_timestamps):
A = np.zeros(3)
dom = tiledb.Domain(tiledb.Dim(domain=(0, 2), tile=3, dtype=np.int64))
att = tiledb.Attr(dtype=A.dtype)
schema = tiledb.ArraySchema(domain=dom, attrs=(att,))
tiledb.DenseArray.create(self.path("foo"), schema)
# write
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = A
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
self.assertEqual(T[0], 0)
self.assertEqual(T[1], 0)
self.assertEqual(T[2], 0)
if use_timestamps:
# sleep 200ms and write
time.sleep(0.2)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[0:1] = 1
if use_timestamps:
# sleep 200ms and write
time.sleep(0.2)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[1:2] = 2
frags = tiledb.array_fragments(self.path("foo"))
# timestamps are in the form of (start, end) for each fragment, with start == end,
# as we are not dealing with consolidated fragments. Let's simply read from 0 to the end timestamp.
read_timestamps = [(0, frag.timestamp_range[1]) for frag in frags]
# read at first timestamp
with tiledb.DenseArray(
self.path("foo"), timestamp=read_timestamps[0], mode="r"
) as T:
self.assertEqual(T[0], 0)
self.assertEqual(T[1], 0)
self.assertEqual(T[2], 0)
# read at second timestamp
with tiledb.DenseArray(
self.path("foo"), timestamp=read_timestamps[1], mode="r"
) as T:
self.assertEqual(T[0], 1)
self.assertEqual(T[1], 0)
self.assertEqual(T[2], 0)
# read at third timestamp
with tiledb.DenseArray(
self.path("foo"), timestamp=read_timestamps[2], mode="r"
) as T:
self.assertEqual(T[0], 1)
self.assertEqual(T[1], 2)
self.assertEqual(T[2], 0)
@pytest.mark.parametrize("use_timestamps", [True, False])
def test_open_timestamp_range(self, use_timestamps):
A = np.zeros(3)
path = self.path("open_timestamp_range")
dom = tiledb.Domain(tiledb.Dim(domain=(0, 2), tile=3, dtype=np.int64))
att = tiledb.Attr(dtype=A.dtype)
schema = tiledb.ArraySchema(domain=dom, attrs=(att,))
tiledb.DenseArray.create(path, schema)
# write
if use_timestamps:
with tiledb.DenseArray(path, mode="w", timestamp=1) as T:
T[:] = A * 1
with tiledb.DenseArray(path, mode="w", timestamp=2) as T:
T[:] = A * 2
with tiledb.DenseArray(path, mode="w", timestamp=3) as T:
T[:] = A * 3
with tiledb.DenseArray(path, mode="w", timestamp=4) as T:
T[:] = A * 4
else:
with tiledb.DenseArray(path, mode="w") as T:
T[:] = A * 1
T[:] = A * 2
T[:] = A * 3
T[:] = A * 4
def assert_ts(timestamp, result):
with tiledb.DenseArray(path, mode="r", timestamp=timestamp) as T:
assert_array_equal(T, result)
timestamps = [t[0] for t in tiledb.array_fragments(path).timestamp_range]
assert_ts(0, A * np.nan)
assert_ts(timestamps[0], A * 1)
assert_ts(timestamps[1], A * 2)
assert_ts(timestamps[2], A * 3)
assert_ts((timestamps[0], timestamps[1]), A * 2)
assert_ts((0, timestamps[2]), A * 3)
assert_ts((timestamps[0], timestamps[2]), A * 3)
assert_ts((timestamps[1], timestamps[2]), A * 3)
assert_ts((timestamps[1], timestamps[3]), A * 3)
assert_ts((None, timestamps[1]), A * 2)
assert_ts((None, timestamps[2]), A * 3)
assert_ts((timestamps[1], None), A * 3)
assert_ts((timestamps[2], None), A * 3)
assert_ts((timestamps[2], None), A * 3)
def test_open_attr(self):
uri = self.path("test_open_attr")
schema = tiledb.ArraySchema(
domain=tiledb.Domain(
tiledb.Dim(name="dim0", dtype=np.uint32, domain=(1, 4))
),
attrs=(
tiledb.Attr(name="x", dtype=np.int32),
tiledb.Attr(name="y", dtype=np.int32),
),
)
tiledb.Array.create(uri, schema)
with tiledb.open(uri, mode="w") as A:
A[:] = {"x": np.array((1, 2, 3, 4)), "y": np.array((5, 6, 7, 8))}
with self.assertRaises(KeyError):
tiledb.open(uri, attr="z")
with self.assertRaises(KeyError):
tiledb.open(uri, attr="dim0")
with tiledb.open(uri, attr="x") as A:
assert_array_equal(A[:], np.array((1, 2, 3, 4)))
assert list(A.multi_index[:].keys()) == ["x"]
def test_ncell_attributes(self):
dom = tiledb.Domain(tiledb.Dim(domain=(0, 9), tile=10, dtype=int))
attr = tiledb.Attr(dtype=[("", np.int32), ("", np.int32), ("", np.int32)])
schema = tiledb.ArraySchema(domain=dom, attrs=(attr,))
tiledb.DenseArray.create(self.path("foo"), schema)
A = np.ones((10,), dtype=[("", np.int32), ("", np.int32), ("", np.int32)])
self.assertEqual(A.dtype, attr.dtype)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = A
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
assert_array_equal(A, T[:])
assert_array_equal(A[:5], T[:5])
def test_complex_attributes(self):
dom = tiledb.Domain(tiledb.Dim(domain=(0, 9), tile=10, dtype=int))
attr = tiledb.Attr(dtype=np.complex64)
schema = tiledb.ArraySchema(domain=dom, attrs=(attr,))
tiledb.DenseArray.create(self.path("foo"), schema)
A = np.random.rand(20).astype(np.float32).view(dtype=np.complex64)
self.assertEqual(schema, tiledb.schema_like(A, dim_dtype=int))
self.assertEqual(A.dtype, attr.dtype)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = A
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
assert_array_equal(A, T[:])
assert_array_equal(A[:5], T[:5])
def test_multiple_attributes(self):
dom = tiledb.Domain(
tiledb.Dim(domain=(0, 1), tile=1, dtype=np.int64),
tiledb.Dim(domain=(0, 3), tile=4, dtype=np.int64),
)
attr_int = tiledb.Attr("ints", dtype=int)
attr_float = tiledb.Attr("floats", dtype=float)
schema = tiledb.ArraySchema(domain=dom, attrs=(attr_int, attr_float))
tiledb.DenseArray.create(self.path("foo"), schema)
V_ints = np.array([[0, 1, 2, 3], [4, 6, 7, 5]])
V_floats = np.array([[0.0, 1.0, 2.0, 3.0], [4.0, 6.0, 7.0, 5.0]])
V = {"ints": V_ints, "floats": V_floats}
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = V
# check setting attribute in different order from Attr definition
# https://github.com/TileDB-Inc/TileDB-Py/issues/299
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = V
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
R = T[:]
assert_array_equal(V["ints"], R["ints"])
assert_array_equal(V["floats"], R["floats"])
R = T.query(attrs=("ints",))[1:3]
assert_array_equal(V["ints"][1:3], R["ints"])
R = T.query(attrs=("floats",), order="F")[:]
self.assertTrue(R["floats"].flags.f_contiguous)
R = T.query(attrs=("ints",), coords=True)[0, 0:3]
self.assertTrue("__dim_0" in R)
self.assertTrue("__dim_1" in R)
assert_array_equal(R["__dim_0"], np.array([0, 0, 0]))
assert_array_equal(R["__dim_1"], np.array([0, 1, 2]))
# Global order returns results as a linear buffer
R = T.query(attrs=("ints",), order="G")[:]
self.assertEqual(R["ints"].shape, (8,))
with self.assertRaises(tiledb.TileDBError):
T.query(attrs=("unknown",))[:]
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
# check error ncells length
V["ints"] = V["ints"][1:2].copy()
with self.assertRaises(tiledb.TileDBError):
T[:] = V
# check error attribute does not exist
V["foo"] = V["ints"].astype(np.int8)
with self.assertRaises(tiledb.TileDBError):
T[:] = V
def test_array_2d_s1(self):
# This array is currently read back with dtype object
A = np.array([["A", "B"], ["C", ""]], dtype="S")
uri = self.path()
dom = tiledb.Domain(
tiledb.Dim(name="rows", domain=(0, 1), tile=2, dtype=np.int64),
tiledb.Dim(name="cols", domain=(0, 1), tile=2, dtype=np.int64),
)
schema = tiledb.ArraySchema(
domain=dom, sparse=False, attrs=[tiledb.Attr(name="a", dtype="S")]
)
tiledb.DenseArray.create(uri, schema)
with tiledb.DenseArray(uri, mode="w") as T:
T[...] = A
with tiledb.DenseArray(uri) as T:
assert_array_equal(A, T)
res = T.multi_index[(0, 1), (0, 1)]["a"]
assert_array_equal(A, res)
def test_nd_roundtrip(self):
dim_set = np.int64([3 + x % 2 for x in range(2, 12)])
for i, last in enumerate(range(2, len(dim_set))):
dims = dim_set[:last]
data = np.random.rand(*dims).astype("int32")
with tiledb.from_numpy(self.path(f"nd_roundtrip{i}"), data) as A:
assert_array_equal(data, A[:])
def test_array_2d_s3_mixed(self):
# This array is currently read back with dtype object
A = np.array([["AAA", "B"], ["AB", "C"]], dtype="S3")
uri = self.path()
dom = tiledb.Domain(
tiledb.Dim(name="rows", domain=(0, 1), tile=2, dtype=np.int64),
tiledb.Dim(name="cols", domain=(0, 1), tile=2, dtype=np.int64),
)
schema = tiledb.ArraySchema(
domain=dom, sparse=False, attrs=[tiledb.Attr(name="a", dtype="S3")]
)
tiledb.DenseArray.create(uri, schema)
with tiledb.DenseArray(uri, mode="w") as T:
T[...] = A
with tiledb.DenseArray(uri) as T:
assert_array_equal(A, T)
res = T.multi_index[(0, 1), (0, 1)]["a"]
assert_array_equal(A, res)
def test_incomplete_dense(self):
path = self.path("incomplete_dense")
# create 10 MB array
data = np.arange(1310720, dtype=np.int64)
# if `tile` is not set, it defaults to the full array and we
# only read 8 bytes at a time.
use_tile = 131072
# use_tile = None
with tiledb.from_numpy(path, data, tile=use_tile) as A:
pass
# create context with 1 MB memory budget (2 MB total, 1 MB usable)
config = tiledb.Config(
{"sm.memory_budget": 2 * 1024**2, "py.init_buffer_bytes": 1024**2}
)
self.assertEqual(config["py.init_buffer_bytes"], str(1024**2))
# TODO would be good to check repeat count here. Not currently exposed by retry loop.
with tiledb.DenseArray(path, ctx=tiledb.Ctx(config)) as A:
res_mr = A.multi_index[slice(0, len(data) - 1)]
assert_array_equal(res_mr[""], data)
res_idx = A[:]
assert_array_equal(res_idx, data)
if has_pandas():
df = A.df[:]
assert_array_equal(df[""], data)
def test_written_fragment_info(self):
uri = self.path("test_written_fragment_info")
dom = tiledb.Domain(tiledb.Dim(domain=(0, 9), tile=10, dtype=np.int64))
att = tiledb.Attr(dtype=np.int64)
schema = tiledb.ArraySchema(domain=dom, attrs=(att,))
tiledb.DenseArray.create(uri, schema)
with tiledb.DenseArray(uri, mode="w") as T:
T[:] = np.arange(0, 10, dtype=np.int64)
self.assertTrue(T.last_write_info is not None)
self.assertTrue(len(T.last_write_info.keys()) == 1)
t_w1, t_w2 = list(T.last_write_info.values())[0]
self.assertTrue(t_w1 > 0)
self.assertTrue(t_w2 > 0)
def test_missing_schema_error(self):
uri = self.path("test_missing_schema_error")
dom = tiledb.Domain(tiledb.Dim(domain=(0, 9), tile=10, dtype=np.int64))
att = tiledb.Attr(dtype=np.int64)
schema = tiledb.ArraySchema(domain=dom, attrs=(att,))
tiledb.DenseArray.create(uri, schema)
with tiledb.DenseArray(uri, mode="w") as T:
T[:] = np.arange(0, 10, dtype=np.int64)
if tiledb.libtiledb.version() < (2, 4):
tiledb.VFS().remove_file(os.path.join(uri, "__array_schema.tdb"))
else:
tiledb.VFS().remove_dir(os.path.join(uri, "__schema"))
# new ctx is required running against S3 because otherwise the schema
# will simply be read from the cache.
with tiledb.scope_ctx():
with self.assertRaises(tiledb.TileDBError):
tiledb.DenseArray(uri)
@pytest.mark.xfail(
tiledb.libtiledb.version() >= (2, 5),
reason="Skip sparse_write_to_dense with libtiledb 2.5+",
)
def test_sparse_write_to_dense(self):
class AssignAndCheck:
def __init__(self, outer, *shape):
self.outer = outer
self.shape = shape
def __setitem__(self, s, v):
A = np.random.rand(*self.shape)
uri = self.outer.path(
f"sparse_write_to_dense{random.randint(0,np.uint64(-1))}"
)
tiledb.from_numpy(uri, A).close()
with tiledb.open(uri, "w") as B:
B[s] = v
A[s] = v
with tiledb.open(uri) as B:
assert_array_equal(A, B[:])
D = AssignAndCheck(self, 5, 5)
with pytest.warns(
DeprecationWarning, match="Sparse writes to dense arrays is deprecated"
):
D[np.array([1, 2]), np.array([0, 0])] = np.array([0, 2])
@pytest.mark.parametrize("use_timestamps", [True, False])
def test_reopen_dense_array(self, use_timestamps):
uri = self.path("test_reopen_dense_array")
dom = tiledb.Domain(tiledb.Dim(domain=(0, 9), tile=10, dtype=np.int64))
att = tiledb.Attr(dtype=np.int64)
schema = tiledb.ArraySchema(domain=dom, attrs=(att,))
tiledb.DenseArray.create(uri, schema)
data = np.arange(0, 10, dtype=np.int64)
if use_timestamps:
with tiledb.DenseArray(uri, mode="w", timestamp=1) as T:
T[:] = data
with tiledb.DenseArray(uri, mode="w", timestamp=2) as T:
T[:] = data * 2
else:
with tiledb.DenseArray(uri, mode="w") as T:
T[:] = data
T[:] = data * 2
timestamps = [t[0] for t in tiledb.array_fragments(uri).timestamp_range]
T = tiledb.DenseArray(uri, mode="r", timestamp=timestamps[0])
assert_array_equal(T[:], data)
T.reopen()
assert_array_equal(T[:], data * 2)
T.close()
def test_data_begins_with_null_chars(self):
path = self.path("test_data_begins_with_null_chars")
data = np.array(["", "", "", "a", "", "", "", "", "", "b"], dtype=np.str_)
dom = tiledb.Domain(tiledb.Dim(domain=(1, len(data)), tile=len(data)))
att = tiledb.Attr(dtype=np.str_, var=True)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.Array.create(path, schema)
with tiledb.open(path, mode="w") as T:
T[:] = data
with tiledb.open(path, mode="r") as T:
assert_array_equal(data, T[:])
def test_match_numpy_schema_dimensions(self):
path = self.path("test_match_numpy_schema_dimensions")
dom = tiledb.Domain(
tiledb.Dim(name="dim_0", domain=(1, 5), dtype=np.int64),
tiledb.Dim(name="dim_1", domain=(1, 10), dtype=np.int64),
tiledb.Dim(name="dim_2", domain=(1, 20), dtype=np.int64),
)
att = tiledb.Attr(name="a", dtype=np.int64)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.Array.create(path, schema)
with tiledb.open(path, "w") as A:
with self.assertRaises(ValueError):
A[:] = np.zeros((10, 20, 5))
A[:] = np.zeros((5, 10, 20))
class TestVarlen(DiskTestCase):
def test_varlen_write_bytes(self):
A = np.array(
[
"aa",
"bbb",
"ccccc",
"ddddddddddddddddddddd",
"ee",
"ffffff",
"g",
"hhhhhhhhhh",
],
dtype=bytes,
)
# basic write
dom = tiledb.Domain(tiledb.Dim(domain=(1, len(A)), tile=len(A)))
att = tiledb.Attr(dtype=np.bytes_)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.DenseArray.create(self.path("foo"), schema)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = A
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
assert_array_equal(A[:], T[:])
assert_array_equal(A, T.multi_index[1 : len(A)][""])
def test_varlen_sparse_all_empty_strings(self):
A = np.array(["", "", "", "", ""], dtype=object)
dim_len = len(A)
uri = self.path("varlen_all_empty_strings")
dom = tiledb.Domain(tiledb.Dim(domain=(1, dim_len), tile=dim_len))
att = tiledb.Attr(name="a1", dtype=np.str_, var=True)
schema = tiledb.ArraySchema(dom, (att,), sparse=True)
tiledb.Array.create(uri, schema)
with tiledb.open(uri, mode="w") as T:
T[np.arange(1, dim_len + 1)] = {"a1": A}
with tiledb.open(uri, mode="r") as T:
# check interior range
assert_array_equal(A[1:-1], T[2:-1]["a1"])
assert_array_equal(A[1:-1], T.multi_index[2 : dim_len - 1]["a1"])
def test_varlen_write_unicode(self):
A = np.array(
[
"aa",
"bbb",
"ccccc",
"ddddddddddddddddddddd",
"ee",
"ffffff",
"g",
"",
"hhhhhhhhhh",
],
dtype=np.str_,
)
# basic write
dom = tiledb.Domain(tiledb.Dim(domain=(1, len(A)), tile=len(A)))
att = tiledb.Attr(dtype=np.str_, var=True)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.DenseArray.create(self.path("foo"), schema)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = A
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
assert_array_equal(A[:], T[:])
def test_varlen_write_floats(self):
# Generates 8 variable-length float64 subarrays (subarray len and content are randomized)
A = np.array(
[np.random.rand(x) for x in np.random.randint(1, 12, 8)], dtype=object
)
# basic write
dom = tiledb.Domain(tiledb.Dim(domain=(1, len(A)), tile=len(A)))
att = tiledb.Attr(dtype=np.float64, var=True)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.DenseArray.create(self.path("foo"), schema)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = A
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
T_ = T[:]
# TODO/note: the return is a 0-element array.
assert_array_equal(A[0], T[1][()])
assert_array_equal(A[-1], T[-1][()])
self.assertEqual(len(A), len(T_))
# can't use assert_array_equal w/ object array
self.assertTrue(all(np.array_equal(x, A[i]) for i, x in enumerate(T_)))
def test_varlen_write_floats_2d(self):
A = np.array(
[np.random.rand(x) for x in np.arange(1, 10)], dtype=object
).reshape(3, 3)
# basic write
dom = tiledb.Domain(
tiledb.Dim(domain=(1, 3), tile=len(A)),
tiledb.Dim(domain=(1, 3), tile=len(A)),
)
att = tiledb.Attr(dtype=np.float64, var=True)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.DenseArray.create(self.path("foo"), schema)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = A
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
T_ = T[:]
self.assertEqual(len(A), len(T_))
# can't use assert_array_equal w/ object array
self.assertTrue(
np.all(
[np.array_equal(A.flat[i], T[:].flat[i]) for i in np.arange(0, 9)]
)
)
def test_varlen_write_int_subarray(self):
A = np.array(
list(
map(
lambda x: np.array(x, dtype=np.uint64),
[np.arange(i, 2 * i + 1) for i in np.arange(0, 16)],
)
),
dtype="O",
).reshape(4, 4)
uri = self.path("test_varlen_write_int_subarray")
dom = tiledb.Domain(
tiledb.Dim(domain=(0, 3), tile=len(A)),
tiledb.Dim(domain=(0, 3), tile=len(A)),
)
att = tiledb.Attr(dtype=np.uint64, var=True)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.DenseArray.create(uri, schema)
# NumPy forces single-element object arrays into a contiguous layout
# so we alternate the size to get a consistent baseline array.
A_onestwos = np.array(
list(
map(
lambda x: np.array(x, dtype=np.uint64),
list([(1,) if x % 2 == 0 else (1, 2) for x in range(16)]),
)
),
dtype=np.dtype("O"),
).reshape(4, 4)
with tiledb.open(uri, "w") as T:
T[:] = A_onestwos
with tiledb.open(uri, "w") as T:
T[1:3, 1:3] = A[1:3, 1:3]
A_assigned = A_onestwos.copy()
A_assigned[1:3, 1:3] = A[1:3, 1:3]
with tiledb.open(uri) as T:
assert_subarrays_equal(A_assigned, T[:])
def test_varlen_write_fixedbytes(self):
# The actual dtype of this array is 'S21'
A = np.array(
[
"aa",
"bbb",
"ccccc",
"ddddddddddddddddddddd",
"ee",
"ffffff",
"g",
"hhhhhhhhhh",
],
dtype=np.dtype("S"),
)
# basic write
dom = tiledb.Domain(tiledb.Dim(domain=(1, len(A)), tile=len(A)))
att = tiledb.Attr(dtype=np.bytes_)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.DenseArray.create(self.path("foo"), schema)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = A
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
assert_array_equal(A[:], T[:])
def test_varlen_write_fixedunicode(self):
A = np.array(
[
"aa",
"bbb",
"ccccc",
"ddddddddddddddddddddd",
"ee",
"ffffff",
"",
"g",
"hhhhhhhhhh",
],
dtype=np.dtype("U"),
)
# basic write
dom = tiledb.Domain(tiledb.Dim(domain=(1, len(A)), tile=len(A)))
att = tiledb.Attr(dtype=np.str_)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.DenseArray.create(self.path("foo"), schema)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = A
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
assert_array_equal(A[:], T[:])
def test_varlen_write_ints(self):
A = np.array(
[
np.uint64(np.random.randint(0, pow(10, 6), x))
for x in np.random.randint(1, 12, 8)
],
dtype=object,
)
# basic write
dom = tiledb.Domain(tiledb.Dim(domain=(1, len(A)), tile=len(A)))
att = tiledb.Attr(dtype=np.int64, var=True)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.DenseArray.create(self.path("foo"), schema)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = A
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
T_ = T[:]
self.assertEqual(len(A), len(T))
# can't use assert_array_equal w/ object array
self.assertTrue(all(np.array_equal(x, A[i]) for i, x in enumerate(T_)))
def test_varlen_wrong_domain(self):
A = np.array(
[
"aa",
"bbb",
"ccccc",
"ddddddddddddddddddddd",
"ee",
"ffffff",
"g",
"hhhhhhhhhh",
]
)
dom = tiledb.Domain(tiledb.Dim(domain=(1, 3), tile=3))
att = tiledb.Attr(dtype=np.bytes_)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.DenseArray.create(self.path("foo"), schema)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
with self.assertRaises(ValueError):
T[:] = A
def test_array_varlen_mismatched(self):
# Test that we raise a TypeError when passing a heterogeneous object array.
A = np.array([b"aa", b"bbb", b"cccc", np.uint64([1, 3, 4])], dtype=object)
dom = tiledb.Domain(tiledb.Dim(domain=(0, 3), tile=4))
att = tiledb.Attr(dtype=np.bytes_, var=True)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.DenseArray.create(self.path("foo"), schema)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
with self.assertRaises(TypeError):
T[:] = A
def test_array_varlen_2d_s_fixed(self):
A = np.array(
[["AAAAAAAAAa", "BBB"], ["ACCC", "BBBCBCBCBCCCBBCBCBCCBC"]], dtype="S"
)
uri = self.path("varlen_2d_s_fixed")
dom = tiledb.Domain(
tiledb.Dim(name="rows", domain=(0, 1), tile=2, dtype=np.int64),
tiledb.Dim(name="cols", domain=(0, 1), tile=2, dtype=np.int64),
)
schema = tiledb.ArraySchema(
domain=dom, sparse=False, attrs=[tiledb.Attr(name="a", dtype="S", var=True)]
)
tiledb.DenseArray.create(uri, schema)
with tiledb.DenseArray(uri, mode="w") as T:
T[...] = A
with tiledb.DenseArray(uri) as T:
assert_array_equal(A, T)
class TestSparseArray(DiskTestCase):
@pytest.mark.xfail
def test_simple_1d_sparse_vector(self):
dom = tiledb.Domain(tiledb.Dim(domain=(0, 3), tile=4, dtype=int))
att = tiledb.Attr(dtype=int)
schema = tiledb.ArraySchema(domain=dom, attrs=(att,), sparse=True)
tiledb.SparseArray.create(self.path("foo"), schema)
values = np.array([3, 4])
with tiledb.SparseArray(self.path("foo"), mode="w") as T:
T[[1, 2]] = values
with tiledb.SparseArray(self.path("foo"), mode="r") as T:
assert_array_equal(T[[1, 2]], values)
@pytest.mark.xfail
def test_simple_2d_sparse_vector(self):
attr = tiledb.Attr(dtype=float)
dom = tiledb.Domain(
tiledb.Dim(domain=(0, 3), tile=4, dtype=int),
tiledb.Dim(domain=(0, 3), tile=4, dtype=int),
)
schema = tiledb.ArraySchema(domain=dom, attrs=(attr,), sparse=True)
tiledb.SparseArray.create(self.path("foo"), schema)
values = np.array([3, 4], dtype=float)
with tiledb.SparseArray(self.path("foo"), mode="w") as T:
T[[1, 2], [1, 2]] = values
with tiledb.SparseArray(self.path("foo"), mode="r") as T:
assert_array_equal(T[[1, 2], [1, 2]], values)
def test_simple3d_sparse_vector(self):
uri = self.path("simple3d_sparse_vector")
dom = tiledb.Domain(
tiledb.Dim("x", domain=(0, 3), tile=4, dtype=int),
tiledb.Dim("y", domain=(0, 3), tile=4, dtype=int),
tiledb.Dim("z", domain=(0, 3), tile=4, dtype=int),
)
attr = tiledb.Attr(dtype=float)
schema = tiledb.ArraySchema(domain=dom, attrs=(attr,), sparse=True)
tiledb.SparseArray.create(uri, schema)
values = np.array([3, 4], dtype=float)
coords = (1, 2), (1, 2), (1, 2)
with tiledb.SparseArray(uri, mode="w") as T:
T[coords] = values
with tiledb.SparseArray(uri, mode="r") as T:
res = T.multi_index[coords]
assert_array_equal(res[""], values)
assert_array_equal(res["x"], coords[0])
assert_array_equal(res["y"], coords[1])
assert_array_equal(res["z"], coords[2])
@pytest.mark.xfail
def test_sparse_ordered_fp_domain(self):
dom = tiledb.Domain(tiledb.Dim("x", domain=(0.0, 10.0), tile=2.0, dtype=float))
attr = tiledb.Attr(dtype=float)
attr = tiledb.Attr(dtype=float)
schema = tiledb.ArraySchema(domain=dom, attrs=(attr,), sparse=True)
tiledb.SparseArray.create(self.path("foo"), schema)
values = np.array([3.3, 2.7])
with tiledb.SparseArray(self.path("foo"), mode="w") as T:
T[[2.5, 4.2]] = values
with tiledb.SparseArray(self.path("foo"), mode="r") as T:
assert_array_equal(T[[2.5, 4.2]], values)
@pytest.mark.xfail
def test_sparse_unordered_fp_domain(self):
dom = tiledb.Domain(tiledb.Dim("x", domain=(0.0, 10.0), tile=2.0, dtype=float))
attr = tiledb.Attr(dtype=float)
schema = tiledb.ArraySchema(domain=dom, attrs=(attr,), sparse=True)
tiledb.SparseArray.create(self.path("foo"), schema)
values = np.array([3.3, 2.7])
with tiledb.SparseArray(self.path("foo"), mode="w") as T:
T[[4.2, 2.5]] = values
with tiledb.SparseArray(self.path("foo"), mode="r") as T:
assert_array_equal(T[[2.5, 4.2]], values[::-1])
@pytest.mark.xfail
def test_multiple_attributes(self):
uri = self.path()
dom = tiledb.Domain(
tiledb.Dim(domain=(1, 10), tile=10, dtype=int),
tiledb.Dim(domain=(1, 10), tile=10, dtype=int),
)
attr_int = tiledb.Attr("ints", dtype=int)
attr_float = tiledb.Attr("floats", dtype="float")
schema = tiledb.ArraySchema(
domain=dom, attrs=(attr_int, attr_float), sparse=True
)
tiledb.SparseArray.create(self.path("foo"), schema)
IJ = (np.array([1, 1, 1, 2, 3, 3, 3, 4]), np.array([1, 2, 4, 3, 1, 6, 7, 5]))
V_ints = np.array([0, 1, 2, 3, 4, 6, 7, 5])
V_floats = np.array([0.0, 1.0, 2.0, 3.0, 4.0, 6.0, 7.0, 5.0])
V = {"ints": V_ints, "floats": V_floats}
with tiledb.SparseArray(uri, mode="w") as T:
T[IJ] = V
with tiledb.SparseArray(uri, mode="r") as T:
R = T[IJ]
assert_array_equal(V["ints"], R["ints"])
assert_array_equal(V["floats"], R["floats"])
# check error attribute does not exist
# TODO: should this be an attribute error?
with tiledb.SparseArray(uri, mode="w") as T:
V["foo"] = V["ints"].astype(np.int8)
with self.assertRaises(tiledb.TileDBError):
T[IJ] = V
# check error ncells length
V["ints"] = V["ints"][1:2].copy()
with self.assertRaises(AttributeError):
T[IJ] = V
def test_query_real_multi_index(self, fx_sparse_cell_order):
uri = self.path("query_real_multi_index")
dom = tiledb.Domain(
tiledb.Dim("x", domain=(-10.0, 10.0), tile=2.0, dtype=float)
)
attr = tiledb.Attr("a", dtype=np.float32)
schema = tiledb.ArraySchema(
domain=dom, attrs=(attr,), sparse=True, cell_order=fx_sparse_cell_order
)
tiledb.SparseArray.create(uri, schema)
values = np.array([3.3, 2.7])
with tiledb.SparseArray(uri, mode="w") as T:
T[[2.5, 4.2]] = values
with tiledb.SparseArray(uri, mode="r") as T:
assert_array_equal(
T.query(coords=True).multi_index[-10.0 : np.nextafter(4.2, 0)]["a"],
np.float32(3.3),
)
assert_array_equal(
T.query(coords=True).multi_index[-10.0 : np.nextafter(4.2, 0)]["x"],
np.float32([2.5]),
)
assert_array_equal(
T.query(coords=False).multi_index[-10.0:5.0]["a"],
np.float32([3.3, 2.7]),
)
self.assertTrue(
"coords" not in T.query(coords=False).multi_index[-10.0:5.0]
)
@pytest.mark.skipif(not has_pandas(), reason="pandas>=1.0,<3.0 not installed")
@pytest.mark.parametrize("dtype", ["u1", "u2", "u4", "u8", "i1", "i2", "i4", "i8"])
def test_sparse_index_dtypes(self, dtype):
path = self.path()
dtype_min, dtype_max = DataType.from_numpy(dtype).domain
dim = tiledb.Dim("d0", domain=(dtype_min, dtype_max - 1), dtype=dtype, tile=1)
attr = tiledb.Attr("attr", dtype=dtype)
schema = tiledb.ArraySchema(tiledb.Domain(dim), [attr], sparse=True)
tiledb.SparseArray.create(path, schema)
data = np.arange(0, 3).astype(dtype)
with tiledb.open(path, "w") as A:
A[data] = data
with tiledb.open(path) as B:
assert_array_equal(B[:]["attr"], data)
assert B[data[0]]["attr"] == data[0]
assert B[data[1]]["attr"] == data[1]
assert B.multi_index[data[0]]["attr"] == data[0]
@pytest.mark.skipif(not has_pandas(), reason="pandas>=1.0,<3.0 not installed")
@pytest.mark.skipif(
tiledb.libtiledb.version() < (2, 10),
reason="TILEDB_BOOL introduced in libtiledb 2.10",
)
def test_sparse_index_bool(self):
path = self.path()
data = np.random.randint(0, 1, 10, dtype=bool)
dom = tiledb.Domain(tiledb.Dim("d0", domain=(1, 10), dtype=int))
attr = tiledb.Attr("attr", dtype=np.bool_)
schema = tiledb.ArraySchema(
domain=dom, attrs=(attr,), sparse=True, allows_duplicates=True
)
tiledb.SparseArray.create(path, schema)
with tiledb.open(path, "w") as A:
A[np.arange(1, 11)] = data
with tiledb.open(path) as B:
assert_array_equal(B[:]["attr"], data)
assert_array_equal(B.multi_index[:]["attr"], data)
def test_query_real_exact(self, fx_sparse_cell_order):
"""
Test and demo of querying at floating point representable boundaries
Concise representation of expected behavior:
c0,c1,c2 = [3.0100000000000002, 3.0100000000000007, 3.010000000000001]
values = [1,2,3]
[c0:c0] -> [1]
[c1:c1] -> [2]
[c2:c2] -> [3]
[c0:c1] -> [1,2]
[c0:c2] -> [1,2,3]
[c0 - nextafter(c0,0) : c0] -> [1]
[c0 - nextafter(c0,0) : c0 - nextafter(c0,0)] -> []
[c2:c2+nextafter(c2)] -> [3]
[c2+nextafter(c2) : c2+nextafter(c2)] -> []
"""
uri = self.path()
dom = tiledb.Domain(
tiledb.Dim("x", domain=(-10.0, 10.0), tile=2.0, dtype=float)
)
attr = tiledb.Attr("", dtype=np.float32)
schema = tiledb.ArraySchema(
domain=dom, attrs=(attr,), sparse=True, cell_order=fx_sparse_cell_order
)
tiledb.SparseArray.create(uri, schema)
c0 = np.nextafter(3.01, 4) # smaller
c1 = np.nextafter(c0, 4)
c2 = np.nextafter(c1, 4) # larger
# for debugging use:
# np.set_printoptions(precision=16, floatmode='maxprec')
# print(c0,c1,c2)
values = np.array([1, 2, 3])
with tiledb.SparseArray(uri, mode="w") as T:
T[[c0, c1, c2]] = values
with tiledb.SparseArray(uri, mode="r") as T:
for i, c in enumerate([c0, c1, c2]):
assert_array_equal(T.query(coords=True).multi_index[c:c][""], values[i])
# test (coord, coord + nextafter)
c0_prev = np.nextafter(c0, 0)
c2_next = np.nextafter(c2, 4)
assert_array_equal(T.query(coords=True).multi_index[c0:c1][""], [1, 2])
assert_array_equal(T.query(coords=True).multi_index[c0:c2][""], [1, 2, 3])
assert_array_equal(T.query(coords=True).multi_index[c2:c2_next][""], 3)
assert_array_equal(T.query(coords=True).multi_index[c0_prev:c0][""], 1)
assert_array_equal(
T.query(coords=True).multi_index[c0_prev:c0_prev][""], []
)
# test (coord + nextafter, coord + nextafter)
assert_array_equal(
T.query(coords=True).multi_index[c2_next:c2_next][""], np.array([])
)
# test (coord - nextafter, coord)
assert_array_equal(
T.query(coords=True).multi_index[c0:c1][""], values[[0, 1]]
)
# test (coord - nextafter, coord + nextafter)
assert_array_equal(
T.query(coords=True).multi_index[c0:c2][""], values[[0, 1, 2]]
)
def test_sparse_query_specified_dim_coords(self, fx_sparse_cell_order):
uri = self.path("sparse_query_specified_dim_coords")
dom = tiledb.Domain(
tiledb.Dim("i", domain=(1, 10), tile=1, dtype=int),
tiledb.Dim("j", domain=(11, 20), tile=1, dtype=int),
)
att = tiledb.Attr("", dtype=int)
schema = tiledb.ArraySchema(
domain=dom, attrs=(att,), sparse=True, cell_order=fx_sparse_cell_order
)
tiledb.SparseArray.create(uri, schema)
i = np.array([1, 1, 1, 2, 3, 3, 3, 4])
j = np.array([11, 12, 14, 13, 11, 16, 17, 15])
with tiledb.SparseArray(uri, mode="w") as A:
A[i, j] = np.array([0, 1, 2, 3, 4, 6, 7, 5])
# data is returned in Hilbert order, so we need to check sorted
with tiledb.SparseArray(uri, mode="r") as A:
Ai = A.query(dims=["i"])[:]
self.assertTrue("i" in Ai)
self.assertFalse("j" in Ai)
assert_unordered_equal(Ai["i"], i, fx_sparse_cell_order == "hilbert")
Aj = A.query(dims=["j"])[:]
self.assertFalse("i" in Aj)
self.assertTrue("j" in Aj)
assert_unordered_equal(Aj["j"], j, fx_sparse_cell_order == "hilbert")
Aij = A.query(dims=["i", "j"])[:]
self.assertTrue("i" in Aij)
self.assertTrue("j" in Aij)
assert_unordered_equal(Aij["i"], i, fx_sparse_cell_order == "hilbert")
assert_unordered_equal(Aij["j"], j, fx_sparse_cell_order == "hilbert")
def test_dense_query_specified_dim_coords(self):
uri = self.path("dense_query_specified_dim_coords")
dom = tiledb.Domain(
tiledb.Dim("i", domain=(1, 3), tile=1, dtype=int),
tiledb.Dim("j", domain=(4, 6), tile=1, dtype=int),
)
att = tiledb.Attr("", dtype=int)
schema = tiledb.ArraySchema(domain=dom, attrs=(att,), sparse=False)
tiledb.Array.create(uri, schema)
with tiledb.open(uri, mode="w") as A:
A[:, :] = np.arange(9).reshape(3, 3)
with tiledb.open(uri, mode="r") as A:
i = np.array([[1, 1, 1], [2, 2, 2], [3, 3, 3]])
j = np.array([[4, 5, 6], [4, 5, 6], [4, 5, 6]])
Ai = A.query(dims=["i"])[:]
self.assertTrue("i" in Ai)
self.assertFalse("j" in Ai)
assert_array_equal(Ai["i"], i)
Aj = A.query(dims=["j"])[:]
self.assertFalse("i" in Aj)
self.assertTrue("j" in Aj)
assert_array_equal(Aj["j"], j)
Aij = A.query(dims=["i", "j"])[:]
self.assertTrue("i" in Aij)
self.assertTrue("j" in Aij)
assert_array_equal(Aij["i"], i)
assert_array_equal(Aij["j"], j)
def test_subarray(self, fx_sparse_cell_order):
dom = tiledb.Domain(tiledb.Dim("x", domain=(1, 10000), tile=100, dtype=int))
att = tiledb.Attr("", dtype=float)
schema = tiledb.ArraySchema(
domain=dom, attrs=(att,), sparse=True, cell_order=fx_sparse_cell_order
)
tiledb.SparseArray.create(self.path("foo"), schema)
with tiledb.SparseArray(self.path("foo"), mode="r") as T:
self.assertIsNone(T.nonempty_domain())
with tiledb.SparseArray(self.path("foo"), mode="w") as T:
T[[50, 60, 100]] = [1.0, 2.0, 3.0]
with tiledb.SparseArray(self.path("foo"), mode="r") as T:
self.assertEqual(((50, 100),), T.nonempty_domain())
# stepped ranges are not supported
with self.assertRaises(IndexError) as idxerr:
T[40:61:5]
assert str(idxerr.value) == "steps are not supported for sparse arrays"
# retrieve just valid coordinates in subarray T[40:60]
assert_array_equal(T[40:61]["x"], [50, 60])
# TODO: dropping coords with one anon value returns just an array
res = T.query(coords=False)[40:61]
assert_array_equal(res[""], [1.0, 2.0])
self.assertEqual(("coords" in res), False)
def test_sparse_bytes(self, fx_sparse_cell_order):
dom = tiledb.Domain(tiledb.Dim("x", domain=(1, 10000), tile=100, dtype=int))
att = tiledb.Attr("", var=True, dtype=np.bytes_)
schema = tiledb.ArraySchema(
domain=dom, attrs=(att,), sparse=True, cell_order=fx_sparse_cell_order
)
tiledb.SparseArray.create(self.path("foo"), schema)
with tiledb.SparseArray(self.path("foo"), mode="r") as T:
self.assertIsNone(T.nonempty_domain())
A = np.array(
[b"aaa", b"bbbbbbbbbbbbbbbbbbbb", b"ccccccccccccccccccccccccc"],
dtype=np.bytes_,
)
with tiledb.SparseArray(self.path("foo"), mode="w") as T:
T[[50, 60, 100]] = A
with tiledb.SparseArray(self.path("foo"), mode="r") as T:
self.assertEqual(((50, 100),), T.nonempty_domain())
# retrieve just valid coordinates in subarray T[40:60]
assert_array_equal(T[40:61]["x"], [50, 60])
# TODO: dropping coords with one anon value returns just an array
res = T.query(coords=False)[40:61]
assert_array_equal(res[""], A[0:2])
self.assertEqual(("coords" in res), False)
# empty sparse varlen result
res = T[1000]
assert_array_equal(res[""], np.array("", dtype="S1"))
assert_array_equal(res["x"], np.array([], dtype=np.int64))
def test_sparse_unicode(self, fx_sparse_cell_order):
dom = tiledb.Domain(tiledb.Dim("x", domain=(1, 10000), tile=100, dtype=int))
att = tiledb.Attr("", var=True, dtype=np.str_)
schema = tiledb.ArraySchema(
domain=dom, attrs=(att,), sparse=True, cell_order=fx_sparse_cell_order
)
tiledb.SparseArray.create(self.path("foo"), schema)
with tiledb.SparseArray(self.path("foo"), mode="r") as T:
self.assertIsNone(T.nonempty_domain())
A = np.array(
[
"1234545lkjalsdfj",
"mnopqrs",
"ijkl",
"gh",
"abcdef",
"aαbββcγγγdδδδδ",
"aαbββc",
"",
"γγγdδδδδ",
],
dtype=object,
)
with tiledb.SparseArray(self.path("foo"), mode="w") as T:
T[[3, 4, 5, 6, 7, 50, 60, 70, 100]] = A
with tiledb.SparseArray(self.path("foo"), mode="r") as T:
self.assertEqual(((3, 100),), T.nonempty_domain())
# retrieve just valid coordinates in subarray T[40:60]
assert_array_equal(T[40:61]["x"], [50, 60])
# TODO: dropping coords with one anon value returns just an array
res = T.query(coords=False)[40:61]
assert_array_equal(res[""], A[5:7])
self.assertEqual(("coords" in res), False)
# empty sparse varlen result
res = T[1000]
assert_array_equal(res[""], np.array("", dtype="U1"))
assert_array_equal(res["x"], np.array([], dtype=np.int64))
def test_sparse_query(self, fx_sparse_cell_order):
uri = self.path("test_sparse_query")
dom = tiledb.Domain(
tiledb.Dim("x", domain=(1, 10000), tile=100, dtype=np.float64)
)
att = tiledb.Attr("", dtype=float)
schema = tiledb.ArraySchema(
domain=dom, attrs=(att,), sparse=True, cell_order=fx_sparse_cell_order
)
tiledb.SparseArray.create(uri, schema)
coords = np.random.uniform(low=1, high=10000, size=100)
data = np.random.rand(100)
with tiledb.SparseArray(uri, mode="w") as T:
T[coords] = data
# Test that TILEDB_UNORDERED works correctly
with tiledb.SparseArray(uri, mode="r") as A:
res = A[1:10001][""] # index past the end here to ensure inclusive result
res = A.multi_index[1:10000][""]
assert_array_equal(np.sort(res), np.sort(data))
res = A.query(order="U").multi_index[1:10000][""]
assert_array_equal(np.sort(res), np.sort(data))
def test_sparse_fixes(self, fx_sparse_cell_order):
uri = self.path("test_sparse_fixes")
# indexing a 1 element item in a sparse array
# (issue directly reported)
# the test here is that the indexing does not raise
dims = (
tiledb.Dim("foo", domain=(0, 6), tile=2),
tiledb.Dim("bar", domain=(0, 6), tile=1),
tiledb.Dim("baz", domain=(0, 100), tile=1),
)
dom = tiledb.Domain(*dims)
att = tiledb.Attr(name="strattr", dtype="S1")
schema = tiledb.ArraySchema(
domain=dom, attrs=(att,), sparse=True, cell_order=fx_sparse_cell_order
)
tiledb.SparseArray.create(uri, schema)
with tiledb.SparseArray(uri) as T:
T[:]
# - test that assigning incompatible value to fixed-len str raises error
# - test that value-conversion error raises exception w/ attr name context
c = np.vstack(
list((x, y, z) for x in range(7) for y in range(7) for z in range(101))
)
with tiledb.SparseArray(uri, "w") as T:
with self.assertRaises(ValueError):
T[c[:, 0], c[:, 1], c[:, 2]] = {"strattr": np.random.rand(7, 7, 101)}
save_exc = list()
try:
T[c[:, 0], c[:, 1], c[:, 2]] = {"strattr": np.random.rand(7, 7, 101)}
except ValueError as e:
save_exc.append(e)
exc = save_exc.pop()
self.assertEqual(
str(exc.__context__),
"Cannot write a string value to non-string typed attribute 'strattr'!",
)
@tiledb.scope_ctx({"sm.check_coord_dups": False})
def test_sparse_fixes_ch1560(self, fx_sparse_cell_order):
uri = self.path("sparse_fixes_ch1560")
schema = tiledb.ArraySchema(
domain=tiledb.Domain(
*[tiledb.Dim(name="id", domain=(1, 5000), tile=25, dtype="int32")]
),
attrs=[
tiledb.Attr(name="a1", dtype="datetime64[s]"),
tiledb.Attr(name="a2", dtype="|S0"),
tiledb.Attr(name="a3", dtype="|S0"),
tiledb.Attr(name="a4", dtype="int32"),
tiledb.Attr(name="a5", dtype="int8"),
tiledb.Attr(name="a6", dtype="int32"),
],
cell_order=fx_sparse_cell_order,
tile_order="row-major",
sparse=True,
)
tiledb.SparseArray.create(uri, schema)
data = OrderedDict(
[
(
"a1",
np.array(
[
"2017-04-01T04:00:00",
"2019-10-01T00:00:00",
"2019-10-01T00:00:00",
"2019-10-01T00:00:00",
],
dtype="datetime64[s]",
),
),
("a2", [b"Bus", b"The RIDE", b"The RIDE", b"The RIDE"]),
("a3", [b"Bus", b"The RIDE", b"The RIDE", b"The RIDE"]),
("a4", np.array([6911721, 138048, 138048, 138048], dtype="int32")),
("a5", np.array([20, 23, 23, 23], dtype="int8")),
("a6", np.array([345586, 6002, 6002, 6002], dtype="int32")),
]
)
with tiledb.open(uri, "w") as A:
A[[1, 462, 462, 462]] = data
with tiledb.open(uri) as A:
res = A[:]
res.pop("id")
for k, v in res.items():
if isinstance(data[k], (np.ndarray, list)):
assert_array_equal(res[k], data[k])
else:
self.assertEqual(res[k], data[k])
def test_sparse_2d_varlen_int(self, fx_sparse_cell_order):
path = self.path("test_sparse_2d_varlen_int")
dtype = np.int32
dom = tiledb.Domain(
tiledb.Dim(domain=(1, 4), tile=2), tiledb.Dim(domain=(1, 4), tile=2)
)
att = tiledb.Attr(dtype=dtype, var=True)
schema = tiledb.ArraySchema(
dom, (att,), sparse=True, cell_order=fx_sparse_cell_order
)
tiledb.SparseArray.create(path, schema)
if tiledb.libtiledb.version() >= (2, 3) and fx_sparse_cell_order == "hilbert":
c1 = np.array([2, 1, 3, 4])
c2 = np.array([1, 2, 3, 4])
else:
c1 = np.array([1, 2, 3, 4])
c2 = np.array([2, 1, 3, 4])
data = np.array(
[
np.array([1, 1], dtype=np.int32),
np.array([2], dtype=np.int32),
np.array([3, 3, 3], dtype=np.int32),
np.array([4], dtype=np.int32),
],
dtype="O",
)
with tiledb.SparseArray(path, "w") as A:
A[c1, c2] = data
with tiledb.SparseArray(path) as A:
res = A[:]
assert_subarrays_equal(res[""], data)
assert_unordered_equal(res["__dim_0"], c1)
assert_unordered_equal(res["__dim_1"], c2)
def test_sparse_mixed_domain_uint_float64(self, fx_sparse_cell_order):
path = self.path("mixed_domain_uint_float64")
dims = [
tiledb.Dim(name="index", domain=(0, 51), tile=11, dtype=np.uint64),
tiledb.Dim(name="dpos", domain=(-100.0, 100.0), tile=10, dtype=np.float64),
]
dom = tiledb.Domain(*dims)
attrs = [tiledb.Attr(name="val", dtype=np.float64)]
schema = tiledb.ArraySchema(
domain=dom, attrs=attrs, sparse=True, cell_order=fx_sparse_cell_order
)
tiledb.SparseArray.create(path, schema)
data = np.random.rand(50, 63)
coords1 = np.repeat(np.arange(0, 50), 63)
coords2 = np.linspace(-100.0, 100.0, num=3150)
with tiledb.open(path, "w") as A:
A[coords1, coords2] = data
# tiledb returns coordinates in sorted order, so we need to check the output
# sorted by the first dim coordinates
sidx = np.argsort(coords1, kind="stable")
coords2_idx = np.tile(np.arange(0, 63), 50)[sidx]
with tiledb.open(path) as A:
res = A[:]
assert_subarrays_equal(
data[coords1[sidx], coords2_idx[sidx]],
res["val"],
fx_sparse_cell_order != "hilbert",
)
a_nonempty = A.nonempty_domain()
self.assertEqual(a_nonempty[0], (0, 49))
self.assertEqual(a_nonempty[1], (-100.0, 100.0))
def test_sparse_string_domain(self, fx_sparse_cell_order):
path = self.path("sparse_string_domain")
dom = tiledb.Domain(tiledb.Dim(name="d", domain=(None, None), dtype=np.bytes_))
att = tiledb.Attr(name="a", dtype=np.int64)
schema = tiledb.ArraySchema(
domain=dom,
attrs=(att,),
sparse=True,
cell_order=fx_sparse_cell_order,
capacity=10000,
)
tiledb.SparseArray.create(path, schema)
data = [1, 2, 3, 4]
coords = [b"aa", b"bbb", b"c", b"dddd"]
with tiledb.open(path, "w") as A:
A[coords] = data
with tiledb.open(path) as A:
ned = A.nonempty_domain()[0]
res = A[ned[0] : ned[1]]
assert_array_equal(res["a"], data)
self.assertEqual(set(res["d"]), set(coords))
self.assertEqual(A.nonempty_domain(), ((b"aa", b"dddd"),))
def test_sparse_string_domain2(self, fx_sparse_cell_order):
path = self.path("sparse_string_domain2")
with self.assertRaises(ValueError):
dims = [
tiledb.Dim(
name="str", domain=(None, None, None), tile=None, dtype=np.bytes_
)
]
dims = [tiledb.Dim(name="str", domain=(None, None), tile=None, dtype=np.bytes_)]
dom = tiledb.Domain(*dims)
attrs = [tiledb.Attr(name="val", dtype=np.float64)]
schema = tiledb.ArraySchema(
domain=dom, attrs=attrs, sparse=True, cell_order=fx_sparse_cell_order
)
tiledb.SparseArray.create(path, schema)
data = np.random.rand(10)
coords = [rand_ascii_bytes(random.randint(5, 50)) for _ in range(10)]
with tiledb.open(path, "w") as A:
A[coords] = data
with tiledb.open(path) as A:
ned = A.nonempty_domain()[0]
res = A[ned[0] : ned[1]]
self.assertTrue(set(res["str"]) == set(coords))
# must check data ordered by coords
assert_array_equal(res["val"], data[np.argsort(coords, kind="stable")])
def test_sparse_mixed_domain(self, fx_sparse_cell_order):
uri = self.path("sparse_mixed_domain")
dims = [
tiledb.Dim(name="p", domain=(-100.0, 100.0), tile=10, dtype=np.float64),
tiledb.Dim(name="str", domain=(None, None), tile=None, dtype=np.bytes_),
]
dom = tiledb.Domain(*dims)
attrs = [tiledb.Attr(name="val", dtype=np.float64)]
schema = tiledb.ArraySchema(
domain=dom, attrs=attrs, sparse=True, cell_order=fx_sparse_cell_order
)
tiledb.SparseArray.create(uri, schema)
nrows = 5
idx_f64 = np.random.rand(nrows)
idx_str = [rand_ascii(5).encode("utf-8") for _ in range(nrows)]
data = np.random.rand(nrows)
with tiledb.SparseArray(uri, "w") as A:
A[idx_f64, idx_str] = {"val": data}
# test heterogeneous dim nonempty_domain
ned_f64 = (np.array(np.min(idx_f64)), np.array(np.max(idx_f64)))
idx_str.sort()
ned_str = idx_str[0], idx_str[-1]
with tiledb.SparseArray(uri, "r") as A:
self.assertEqual(A.nonempty_domain(), (ned_f64, ned_str))
def test_sparse_get_unique_dim_values(self, fx_sparse_cell_order):
uri = self.path("get_non_empty_coords")
dim1 = tiledb.Dim(name="dim1", domain=(None, None), tile=None, dtype=np.bytes_)
dim2 = tiledb.Dim(name="dim2", domain=(0, 1), tile=1, dtype=np.float64)
attr = tiledb.Attr(name="attr", dtype=np.float32)
dom = tiledb.Domain(dim1, dim2)
schema = tiledb.ArraySchema(
domain=dom, sparse=True, cell_order=fx_sparse_cell_order, attrs=[attr]
)
tiledb.Array.create(uri, schema)
with tiledb.open(uri, "w") as A:
A["a1", 0] = 1
A["a1", 0.25] = 2
A["a2", 0.5] = 3
A["a3", 0.25] = 4
with tiledb.open(uri, "r") as A:
self.assertEqual(
A.unique_dim_values(),
OrderedDict(
[("dim1", (b"a1", b"a2", b"a3")), ("dim2", (0.0, 0.25, 0.5))]
),
)
self.assertEqual(A.unique_dim_values("dim1"), (b"a1", b"a2", b"a3"))
self.assertEqual(A.unique_dim_values("dim2"), (0, 0.25, 0.5))
with self.assertRaises(ValueError):
A.unique_dim_values(0)
with self.assertRaises(ValueError):
A.unique_dim_values("dim3")
def test_sparse_write_for_zero_attrs(self):
uri = self.path("test_sparse_write_to_zero_attrs")
dim = tiledb.Dim(name="dim", domain=(0, 9), dtype=np.float64)
schema = tiledb.ArraySchema(domain=tiledb.Domain(dim), sparse=True)
tiledb.Array.create(uri, schema)
coords = [1, 2.0, 3.5]
with tiledb.open(uri, "w") as A:
A[coords] = None
with tiledb.open(uri, "r") as A:
output = A.query()[:]
assert list(output.keys()) == ["dim"]
assert_array_equal(output["dim"][:], coords)
def test_sparse_write_nullable_default(self):
uri = self.path("test_sparse_write_nullable_default")
dim1 = tiledb.Dim(name="d1", dtype="|S0", var=True)
att = tiledb.Attr(name="a1", dtype="<U0", var=True, nullable=True)
schema = tiledb.ArraySchema(
domain=tiledb.Domain(dim1),
attrs=(att,),
sparse=True,
allows_duplicates=False,
)
tiledb.Array.create(uri, schema)
with tiledb.open(uri, "w") as A:
A[["a", "b", "c"]] = np.array(["aaa", "bb", "c"])
if has_pandas():
import pandas as pd
with tiledb.open(uri) as A:
pd._testing.assert_frame_equal(
A.query(dims=False).df[:],
pd.DataFrame({"a1": pd.Series(["aaa", "bb", "c"])}),
)
class TestDenseIndexing(DiskTestCase):
def _test_index(self, A, T, idx):
expected = A[idx]
actual = T[idx]
assert_array_equal(expected, actual)
good_index_1d = [
# single value
42,
-1,
# slices
slice(0, 1050),
slice(50, 150),
slice(0, 2000),
slice(-150, -50),
# TODO: indexing failures
# slice(-2000, 2000),
# slice(0, 0), # empty result
# slice(-1, 0), # empty result
# total selections
slice(None),
Ellipsis,
(),
(Ellipsis, slice(None)),
# slice with step
slice(None),
slice(None, None),
slice(None, None, 1),
slice(None, None, 10),
slice(None, None, 100),
slice(None, None, 1000),
slice(None, None, 10000),
slice(0, 1050),
slice(0, 1050, 1),
slice(0, 1050, 10),
slice(0, 1050, 100),
slice(0, 1050, 1000),
slice(0, 1050, 10000),
slice(1, 31, 3),
slice(1, 31, 30),
slice(1, 31, 300),
slice(81, 121, 3),
slice(81, 121, 30),
slice(81, 121, 300),
slice(50, 150),
slice(50, 150, 1),
slice(50, 150, 10),
# TODO: negative steps
slice(None, None, -1),
slice(None, None, -10),
slice(None, None, -100),
slice(None, None, -1000),
slice(None, None, -10000),
# slice(1050, -1, -1),
# slice(1050, -1, -10),
# slice(1050, -1, -100),
# slice(1050, -1, -1000),
# slice(1050, -1, -10000),
# slice(1050, 0, -1),
# slice(1050, 0, -10),
# slice(1050, 0, -100),
# slice(1050, 0, -1000),
# slice(1050, 0, -10000),
# slice(150, 50, -1),
# slice(150, 50, -10),
# slice(31, 1, -3),
# slice(121, 81, -3),
# slice(-1, 0, -1),
]
bad_index_1d = ["foo", b"xxx", None, (0, 0), (slice(None), slice(None))]
warn_and_bad_index_1d = [2.3, -4.5]
def test_index_1d(self):
A = np.arange(1050, dtype=int)
dom = tiledb.Domain(tiledb.Dim(domain=(0, 1049), tile=100))
att = tiledb.Attr(dtype=int)
schema = tiledb.ArraySchema(domain=dom, attrs=(att,))
tiledb.DenseArray.create(self.path("foo"), schema)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = A
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
for idx in self.good_index_1d:
self._test_index(A, T, idx)
for idx in self.bad_index_1d:
with self.assertRaises(IndexError):
T[idx]
for idx in self.warn_and_bad_index_1d:
with self.assertRaises(IndexError):
T[idx]
good_index_2d = [
# single row
42,
-1,
(42, slice(None)),
(-1, slice(None)),
# single col
(slice(None), 4),
(slice(None), -1),
# row slices
slice(None),
slice(0, 1000),
slice(250, 350),
slice(0, 2000),
slice(-350, -250),
slice(0, 0), # empty result
slice(-1, 0), # empty result
slice(-2000, 0),
slice(-2000, 2000),
# 2D slices
(slice(None), slice(1, 5)),
(slice(250, 350), slice(None)),
(slice(250, 350), slice(1, 5)),
(slice(250, 350), slice(-5, -1)),
(slice(250, 350), slice(-50, 50)),
(slice(250, 350, 10), slice(1, 5)),
(slice(250, 350), slice(1, 5, 2)),
(slice(250, 350, 33), slice(1, 5, 3)),
# total selections
(slice(None), slice(None)),
Ellipsis,
(),
(Ellipsis, slice(None)),
(Ellipsis, slice(None), slice(None)),
# TODO: negative steps
# slice(None, None, -1),
# (slice(None, None, -1), slice(None)),
]
bad_index_2d = [
"foo",
b"xxx",
None,
(0, 0, 0),
(slice(None), slice(None), slice(None)),
]
warn_and_bad_index_2d = [
2.3,
(2.3, slice(None)),
]
def test_index_2d(self):
A = np.arange(10000).reshape((1000, 10))
dom = tiledb.Domain(
tiledb.Dim(domain=(0, 999), tile=100), tiledb.Dim(domain=(0, 9), tile=2)
)
att = tiledb.Attr(dtype=A.dtype)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.DenseArray.create(self.path("foo"), schema)
with tiledb.DenseArray(self.path("foo"), mode="w") as T:
T[:] = A
with tiledb.DenseArray(self.path("foo"), mode="r") as T:
for idx in self.good_index_1d:
self._test_index(A, T, idx)
for idx in self.bad_index_2d:
with self.assertRaises(IndexError):
T[idx]
for idx in self.warn_and_bad_index_2d:
with self.assertRaises(IndexError):
T[idx]
class TestDatetimeSlicing(DiskTestCase):
def test_dense_datetime_vector(self):
uri = self.path("foo_datetime_vector")
# Domain is 10 years, day resolution, one tile per 365 days
dim = tiledb.Dim(
name="d1",
domain=(np.datetime64("2010-01-01"), np.datetime64("2020-01-01")),
tile=np.timedelta64(365, "D"),
dtype=np.datetime64("", "D").dtype,
)
dom = tiledb.Domain(dim)
schema = tiledb.ArraySchema(
domain=dom, attrs=(tiledb.Attr("a1", dtype=np.float64),)
)
tiledb.Array.create(uri, schema)
# Write a few years of data at the beginning using a timedelta object
ndays = 365 * 2
a1_vals = np.random.rand(ndays)
start = np.datetime64("2010-01-01")
# Datetime indexing is inclusive, so a delta of one less
end = start + np.timedelta64(ndays - 1, "D")
with tiledb.DenseArray(uri, "w") as T:
T[start:end] = {"a1": a1_vals}
# Read back data
with tiledb.DenseArray(uri, "r", attr="a1") as T:
assert_array_equal(T[start:end], a1_vals)
# Check nonempty domain
with tiledb.DenseArray(uri, "r") as T:
nonempty = T.nonempty_domain()
d1_nonempty = nonempty[0]
self.assertEqual(d1_nonempty[0].dtype, np.datetime64("", "D"))
self.assertEqual(d1_nonempty[1].dtype, np.datetime64("", "D"))
self.assertTupleEqual(d1_nonempty, (start, end))
# Slice a few days from the middle using two datetimes
with tiledb.DenseArray(uri, "r", attr="a1") as T:
# Slice using datetimes
actual = T[np.datetime64("2010-11-01") : np.datetime64("2011-01-31")]
# Convert datetime interval to integer offset/length into original array
# must be cast to int because float slices are not allowed in NumPy 1.12+
read_offset = int(
(np.datetime64("2010-11-01") - start) / np.timedelta64(1, "D")
)
read_ndays = int(
(np.datetime64("2011-01-31") - np.datetime64("2010-11-01") + 1)
/ np.timedelta64(1, "D")
)
expected = a1_vals[read_offset : read_offset + read_ndays]
assert_array_equal(actual, expected)
# Slice the first year
with tiledb.DenseArray(uri, "r", attr="a1") as T:
actual = T[np.datetime64("2010") : np.datetime64("2011")]
# Convert datetime interval to integer offset/length into original array
read_offset = int(
(np.datetime64("2010-01-01") - start) / np.timedelta64(1, "D")
)
read_ndays = int(
(np.datetime64("2011-01-01") - np.datetime64("2010-01-01") + 1)
/ np.timedelta64(1, "D")
)
expected = a1_vals[read_offset : read_offset + read_ndays]
assert_array_equal(actual, expected)
# Slice open spans
with tiledb.DenseArray(uri, "r", attr="a1") as T:
# Convert datetime interval to integer offset/length into original array
read_offset = int(
(np.datetime64("2010-01-01") - start) / np.timedelta64(1, "D")
)
read_ndays = int(
(np.datetime64("2011-01-31") - np.datetime64("2010-01-01") + 1)
/ np.timedelta64(1, "D")
)
expected = a1_vals[read_offset : read_offset + read_ndays]
# note we only wrote first two years
actual = T.multi_index[np.datetime64("2010-01-01") :]["a1"][:read_ndays]
assert_array_equal(actual, expected)
actual2 = T[np.datetime64("2010-01-01") :][:read_ndays]
assert_array_equal(actual2, expected)
def test_sparse_datetime_vector(self, fx_sparse_cell_order):
uri = self.path("foo_datetime_sparse_vector")
# ns resolution, one tile per second, max domain possible
dim = tiledb.Dim(
name="d1",
domain=(
np.datetime64(0, "ns"),
np.datetime64(int(np.iinfo(np.int64).max) - 1000000000, "ns"),
),
tile=np.timedelta64(1, "s"),
dtype=np.datetime64("", "ns").dtype,
)
self.assertEqual(dim.tile, np.timedelta64("1000000000", "ns"))
dom = tiledb.Domain(dim)
schema = tiledb.ArraySchema(
domain=dom,
sparse=True,
cell_order=fx_sparse_cell_order,
attrs=(tiledb.Attr("a1", dtype=np.float64),),
)
tiledb.Array.create(uri, schema)
# Write 10k cells every 1000 ns starting at time 0
coords = np.datetime64(0, "ns") + np.arange(0, 10000 * 1000, 1000)
a1_vals = np.random.rand(len(coords))
with tiledb.SparseArray(uri, "w") as T:
T[coords] = {"a1": a1_vals}
# Read all
with tiledb.SparseArray(uri, "r") as T:
assert_array_equal(T[:]["a1"], a1_vals)
# Read back first 10 cells
with tiledb.SparseArray(uri, "r") as T:
start = np.datetime64(0, "ns")
vals = T[start : start + np.timedelta64(10000, "ns")]["a1"]
assert_array_equal(vals, a1_vals[0:11])
# Test open ended ranges multi_index
vals2 = T.multi_index[start:]["a1"]
assert_array_equal(vals2, a1_vals)
stop = np.datetime64(int(np.iinfo(np.int64).max) - 1000000000, "ns")
vals3 = T.multi_index[:stop]["a1"]
assert_array_equal(vals3, a1_vals)
def test_datetime_types(self, fx_sparse_cell_order):
units = ["h", "m", "s", "ms", "us", "ns", "ps", "fs"]
for res in units:
uri = self.path("test_datetime_type_" + res)
tmax = 1000
tile = np.timedelta64(1, res)
dim = tiledb.Dim(
name="d1",
domain=(None, None),
tile=tile,
dtype=np.datetime64("", res).dtype,
)
dom = tiledb.Domain(dim)
schema = tiledb.ArraySchema(
domain=dom,
sparse=True,
cell_order=fx_sparse_cell_order,
attrs=(tiledb.Attr("a1", dtype=np.float64),),
)
tiledb.Array.create(uri, schema)
# Write tmax cells every 10 units starting at time 0
coords = np.datetime64(0, res) + np.arange(
0, tmax, 10
) # np.arange(0, 10000 * 1000, 1000)
a1_vals = np.random.rand(len(coords))
with tiledb.SparseArray(uri, "w") as T:
T[coords] = {"a1": a1_vals}
# Read all
with tiledb.SparseArray(uri, "r") as T:
assert_array_equal(T[:]["a1"], a1_vals)
# Read back first 10 cells
with tiledb.SparseArray(uri, "r") as T:
start = np.datetime64(0, res)
vals = T[start : start + np.timedelta64(int(tmax / 10), res)]["a1"]
assert_array_equal(vals, a1_vals[0:11])
class PickleTest(DiskTestCase):
# test that DenseArray and View can be pickled for multiprocess use
# note that the current pickling is by URI and attributes (it is
# not, and likely should not be, a way to serialize array data)
@pytest.mark.parametrize("sparse", [True, False])
def test_pickle_roundtrip(self, sparse):
uri = self.path("test_pickle_roundtrip")
dom = tiledb.Domain(tiledb.Dim(domain=(0, 2), tile=3))
schema = tiledb.ArraySchema(domain=dom, attrs=(tiledb.Attr(""),), sparse=sparse)
tiledb.libtiledb.Array.create(uri, schema)
with tiledb.open(uri, "w") as T:
if sparse:
T[[0, 1, 2]] = np.random.randint(10, size=3)
else:
T[:] = np.random.randint(10, size=3)
with tiledb.open(uri, "r") as T:
with io.BytesIO() as buf:
pickle.dump(T, buf)
buf.seek(0)
with pickle.load(buf) as T2:
if sparse:
assert_array_equal(T[:][""], T2[:][""])
else:
assert_array_equal(T[:], T2[:])
with io.BytesIO() as buf, tiledb.open(uri) as V:
pickle.dump(V, buf)
buf.seek(0)
with pickle.load(buf) as V2:
# make sure anonymous view pickles and round-trips
if sparse:
assert_array_equal(V[:][""], V2[:][""])
else:
assert_array_equal(V[:], V2[:])
@tiledb.scope_ctx({"vfs.s3.region": "kuyper-belt-1", "vfs.max_parallel_ops": "1"})
def test_pickle_with_config(self):
uri = self.path("pickle_config")
T = tiledb.from_numpy(uri, np.random.rand(3, 3))
with io.BytesIO() as buf:
pickle.dump(T, buf)
buf.seek(0)
T2 = pickle.load(buf)
assert_array_equal(T, T2)
self.maxDiff = None
d1 = tiledb.default_ctx().config().dict()
d2 = T2._ctx_().config().dict()
self.assertEqual(d1["vfs.s3.region"], d2["vfs.s3.region"])
self.assertEqual(d1["vfs.max_parallel_ops"], d2["vfs.max_parallel_ops"])
T.close()
T2.close()
@pytest.mark.parametrize("sparse", [True, False])
@pytest.mark.parametrize("use_timestamps", [True, False])
def test_pickle_with_tuple_timestamps(self, sparse, use_timestamps):
A = np.random.randint(10, size=3)
path = self.path("test_pickle_with_tuple_timestamps")
dom = tiledb.Domain(tiledb.Dim(domain=(0, 2), tile=3, dtype=np.int64))
att = tiledb.Attr(dtype=A.dtype)
schema = tiledb.ArraySchema(domain=dom, attrs=(att,), sparse=sparse)
tiledb.libtiledb.Array.create(path, schema)
for ts in range(1, 5):
with tiledb.open(
path, timestamp=ts if use_timestamps else None, mode="w"
) as T:
if sparse:
T[[0, 1, 2]] = A * ts
else:
T[:] = A * ts
timestamps = [t[0] for t in tiledb.array_fragments(path).timestamp_range]
with tiledb.open(path, timestamp=(timestamps[1], timestamps[2]), mode="r") as T:
with io.BytesIO() as buf:
pickle.dump(T, buf)
buf.seek(0)
with pickle.load(buf) as T2:
if sparse:
assert_array_equal(T[:][""], T2[:][""])
else:
assert_array_equal(T[:], T2[:])
assert T2.timestamp_range == (timestamps[1], timestamps[2])
with io.BytesIO() as buf, tiledb.open(
path, timestamp=(timestamps[1], timestamps[2])
) as V:
pickle.dump(V, buf)
buf.seek(0)
with pickle.load(buf) as V2:
# make sure anonymous view pickles and round-trips
if sparse:
assert_array_equal(V[:][""], V2[:][""])
else:
assert_array_equal(V[:], V2[:])
assert V2.timestamp_range == (timestamps[1], timestamps[2])
class ArrayViewTest(DiskTestCase):
def test_view_multiattr(self):
uri = self.path("foo_multiattr")
dom = tiledb.Domain(
tiledb.Dim(domain=(0, 2), tile=3), tiledb.Dim(domain=(0, 2), tile=3)
)
schema = tiledb.ArraySchema(
domain=dom, attrs=(tiledb.Attr(""), tiledb.Attr("named"))
)
tiledb.libtiledb.Array.create(uri, schema)
anon_ar = np.random.rand(3, 3)
named_ar = np.random.rand(3, 3)
with tiledb.DenseArray(uri, "w") as T:
T[:] = {"": anon_ar, "named": named_ar}
with self.assertRaises(KeyError):
T = tiledb.DenseArray(uri, "r", attr="foo111")
with tiledb.DenseArray(uri, "r", attr="named") as T:
assert_array_equal(T, named_ar)
# make sure each attr view can pickle and round-trip
with io.BytesIO() as buf:
pickle.dump(T, buf)
buf.seek(0)
with pickle.load(buf) as T_rt:
assert_array_equal(T, T_rt)
with tiledb.DenseArray(uri, "r", attr="") as T:
assert_array_equal(T, anon_ar)
with io.BytesIO() as buf:
pickle.dump(T, buf)
buf.seek(0)
with pickle.load(buf) as tmp:
assert_array_equal(tmp, anon_ar)
# set subarray on multi-attribute
range_ar = np.arange(0, 9).reshape(3, 3)
with tiledb.DenseArray(uri, "w", attr="named") as V_named:
V_named[1:3, 1:3] = range_ar[1:3, 1:3]
with tiledb.DenseArray(uri, "r", attr="named") as V_named:
assert_array_equal(V_named[1:3, 1:3], range_ar[1:3, 1:3])
class RWTest(DiskTestCase):
def test_read_write(self, capfd):
dom = tiledb.Domain(tiledb.Dim(domain=(0, 2), tile=3))
att = tiledb.Attr(dtype="int64")
schema = tiledb.ArraySchema(domain=dom, attrs=(att,))
tiledb.libtiledb.Array.create(self.path("foo"), schema)
np_array = np.array([1, 2, 3], dtype="int64")
with tiledb.DenseArray(self.path("foo"), mode="w") as arr:
arr.write_direct(np_array)
with tiledb.DenseArray(self.path("foo"), mode="r") as arr:
arr.dump()
assert_captured(capfd, "Array type: dense")
self.assertEqual(arr.nonempty_domain(), ((0, 2),))
self.assertEqual(arr.ndim, np_array.ndim)
assert_array_equal(arr.read_direct(), np_array)
class TestNumpyToArray(DiskTestCase):
def test_to_array0d(self):
# Cannot create 0-dim arrays in TileDB
np_array = np.array(1)
with self.assertRaises(tiledb.TileDBError):
with tiledb.from_numpy(self.path("foo"), np_array):
pass
def test_to_array1d(self):
np_array = np.array([1.0, 2.0, 3.0])
with tiledb.from_numpy(self.path("foo"), np_array) as arr:
assert_array_equal(arr[:], np_array)
def test_to_array2d(self):
np_array = np.ones((100, 100), dtype="i8")
with tiledb.from_numpy(self.path("foo"), np_array) as arr:
assert_array_equal(arr[:], np_array)
def test_to_array3d(self):
np_array = np.ones((1, 1, 1), dtype="i1")
with tiledb.from_numpy(self.path("foo"), np_array) as arr:
assert_array_equal(arr[:], np_array)
def test_bytes_to_array1d(self):
np_array = np.array(
[b"abcdef", b"gh", b"ijkl", b"mnopqrs", b"", b"1234545lkjalsdfj"],
dtype=object,
)
with tiledb.from_numpy(self.path("foo"), np_array) as arr:
assert_array_equal(arr[:], np_array)
with tiledb.DenseArray(self.path("foo")) as arr_reload:
assert_array_equal(arr_reload[:], np_array)
def test_unicode_to_array1d(self):
np_array = np.array(
[
"1234545lkjalsdfj",
"mnopqrs",
"ijkl",
"gh",
"abcdef",
"aαbββcγγγdδδδδ",
"",
'"aαbββc',
"",
"γγγdδδδδ",
],
dtype=object,
)
with tiledb.from_numpy(self.path("foo"), np_array) as arr:
assert_array_equal(arr[:], np_array)
with tiledb.DenseArray(self.path("foo")) as arr_reload:
assert_array_equal(arr_reload[:], np_array)
def test_array_interface(self):
# Tests that __array__ interface works
np_array1 = np.arange(1, 10)
with tiledb.from_numpy(self.path("arr1"), np_array1) as arr1:
assert_array_equal(np.array(arr1), np_array1)
# Test that __array__ interface throws an error when number of attributes > 1
dom = tiledb.Domain(tiledb.Dim(domain=(0, 2), tile=3))
foo = tiledb.Attr("foo", dtype="i8")
bar = tiledb.Attr("bar", dtype="i8")
schema = tiledb.ArraySchema(domain=dom, attrs=(foo, bar))
tiledb.DenseArray.create(self.path("arr2"), schema)
with self.assertRaises(ValueError):
with tiledb.DenseArray(self.path("arr2"), mode="r") as arr2:
np.array(arr2)
def test_array_getindex(self):
# Tests that __getindex__ interface works
np_array = np.arange(1, 10)
with tiledb.from_numpy(self.path("foo"), np_array) as arr:
assert_array_equal(arr[5:10], np_array[5:10])
def test_to_array1d_attr_name(self):
np_array = np.array([1.0, 2.0, 3.0])
with tiledb.from_numpy(self.path("foo"), np_array, attr_name="a") as arr:
assert_array_equal(arr[:]["a"], np_array)
def test_from_numpy_timestamp(self):
path = self.path()
with tiledb.from_numpy(path, np.array([1, 2, 3]), timestamp=10) as A:
pass
with tiledb.open(path, timestamp=(0, 9)) as A:
assert A.nonempty_domain() is None
with tiledb.open(path, timestamp=(10, 10)) as A:
assert A.nonempty_domain() == ((0, 2),)
def test_from_numpy_schema_only(self):
uri = self.path("test_from_numpy_schema_only")
arr1 = np.array([1.0, 2.0, 3.0])
with tiledb.from_numpy(uri, arr1, mode="schema_only") as arr:
assert arr.nonempty_domain() is None
def test_from_numpy_append(self):
uri = self.path("test_from_numpy_append")
arr1 = np.array([1.0, 2.0, 3.0])
with tiledb.from_numpy(uri, arr1, full_domain=True) as A:
assert A.nonempty_domain() == ((0, 2),)
assert_array_equal(A[0:3], arr1)
arr2 = np.array([4.0, 5.0, 6.0])
with tiledb.from_numpy(uri, arr2, mode="append") as A:
assert A.nonempty_domain() == ((0, 5),)
assert_array_equal(A[0:6], np.append(arr1, arr2))
def test_from_numpy_start_idx(self):
uri = self.path("test_from_numpy_start_idx")
arr1 = np.array([1.0, 2.0, 3.0])
with tiledb.from_numpy(uri, arr1) as A:
assert A.nonempty_domain() == ((0, 2),)
assert_array_equal(A[0:3], arr1)
arr2 = np.array([4.0, 5.0, 6.0])
with tiledb.from_numpy(uri, arr2, mode="append", start_idx=0) as A:
assert A.nonempty_domain() == ((0, 2),)
assert_array_equal(A[0:3], arr2)
def test_from_numpy_append_array2d(self):
uri = self.path("test_from_numpy_append_array2d")
arr1 = np.random.rand(10, 5)
with tiledb.from_numpy(uri, arr1, full_domain=True) as A:
assert A.nonempty_domain() == ((0, 9), (0, 4))
assert_array_equal(A[0:10, 0:5], arr1)
# error out if number of dimensions do not match
with self.assertRaises(ValueError):
arr2 = np.random.rand(5)
tiledb.from_numpy(uri, arr2, mode="append")
# error out if number of dimensions do not match
with self.assertRaises(ValueError):
arr2 = np.random.rand(4, 4)
tiledb.from_numpy(uri, arr2, mode="append")
arr2 = np.random.rand(5, 5)
with tiledb.from_numpy(uri, arr2, mode="append") as A:
assert A.nonempty_domain() == ((0, 14), (0, 4))
assert_array_equal(A[0:15, 0:5], np.append(arr1, arr2, axis=0))
@pytest.mark.parametrize("append_dim", (0, 1, 2, 3))
def test_from_numpy_append_array3d(self, append_dim):
uri = self.path("test_from_numpy_append_array3d")
arr1 = np.random.rand(2, 2, 2)
with tiledb.from_numpy(uri, arr1, full_domain=True) as A:
assert A.nonempty_domain() == ((0, 1), (0, 1), (0, 1))
assert_array_equal(A[0:2, 0:2, 0:2], arr1)
arr2 = np.random.rand(2, 2, 2)
# error out if index is out of bounds
if append_dim == 3:
with self.assertRaises(IndexError):
tiledb.from_numpy(uri, arr2, mode="append", append_dim=append_dim)
return
with tiledb.from_numpy(uri, arr2, mode="append", append_dim=append_dim) as A:
if append_dim == 0:
assert A.nonempty_domain() == ((0, 3), (0, 1), (0, 1))
result = A[0:4, 0:2, 0:2]
elif append_dim == 1:
assert A.nonempty_domain() == ((0, 1), (0, 3), (0, 1))
result = A[0:2, 0:4, 0:2]
elif append_dim == 2:
assert A.nonempty_domain() == ((0, 1), (0, 1), (0, 3))
result = A[0:2, 0:2, 0:4]
assert_array_equal(result, np.append(arr1, arr2, axis=append_dim))
@pytest.mark.parametrize("append_dim", (0, 1, 2, 3))
def test_from_numpy_append_array3d_overwrite(self, append_dim):
uri = self.path("test_from_numpy_append_array3d")
arr1 = np.random.rand(2, 2, 2)
with tiledb.from_numpy(uri, arr1) as A:
assert A.nonempty_domain() == ((0, 1), (0, 1), (0, 1))
assert_array_equal(A[0:2, 0:2, 0:2], arr1)
arr2 = np.random.rand(2, 2, 2)
# error out if index is out of bounds
if append_dim == 3:
with self.assertRaises(IndexError):
tiledb.from_numpy(uri, arr2, mode="append", append_dim=append_dim)
return
with tiledb.from_numpy(
uri, arr2, mode="append", append_dim=append_dim, start_idx=0
) as A:
assert A.nonempty_domain() == ((0, 1), (0, 1), (0, 1))
assert_array_equal(A[0:2, 0:2, 0:2], arr2)
@pytest.mark.parametrize("empty_str", ["", b""])
@pytest.mark.parametrize("num_strs", [1, 1000])
def test_from_numpy_empty_str(self, empty_str, num_strs):
uri = self.path("test_from_numpy_empty_str")
np_array = np.asarray([empty_str] * num_strs, dtype="O")
tiledb.from_numpy(uri, np_array)
with tiledb.open(uri, "r") as A:
assert_array_equal(A[:], np_array)
if has_pandas():
assert_array_equal(A.query(use_arrow=True).df[:][""], np_array)
assert_array_equal(A.query(use_arrow=False).df[:][""], np_array)
class ConsolidationTest(DiskTestCase):
def test_array_vacuum(self):
dshape = (0, 19)
num_writes = 10
def create_array(target_path):
dom = tiledb.Domain(tiledb.Dim(domain=dshape, tile=3))
att = tiledb.Attr(dtype="int64")
schema = tiledb.ArraySchema(domain=dom, attrs=(att,))
tiledb.libtiledb.Array.create(target_path, schema)
def write_fragments(target_path):
for i in range(num_writes):
with tiledb.open(target_path, "w") as A:
A[i : dshape[1]] = np.random.rand(dshape[1] - i)
# array #1
path = self.path("test_array_vacuum")
create_array(path)
write_fragments(path)
fi = tiledb.array_fragments(path)
self.assertEqual(len(fi), num_writes)
tiledb.consolidate(path)
tiledb.vacuum(path)
fi = tiledb.array_fragments(path)
self.assertEqual(len(fi), 1)
# array #2
path2 = self.path("test_array_vacuum_fragment_meta")
create_array(path2)
write_fragments(path2)
fi = tiledb.array_fragments(path2)
self.assertEqual(fi.unconsolidated_metadata_num, num_writes)
tiledb.consolidate(
path2, config=tiledb.Config({"sm.consolidation.mode": "fragment_meta"})
)
tiledb.vacuum(path2, config=tiledb.Config({"sm.vacuum.mode": "fragment_meta"}))
fi = tiledb.array_fragments(path2)
self.assertEqual(fi.unconsolidated_metadata_num, 0)
# array #3
path3 = self.path("test_array_vacuum2")
create_array(path3)
write_fragments(path3)
fi = tiledb.array_fragments(path3)
self.assertEqual(fi.unconsolidated_metadata_num, num_writes)
conf = tiledb.Config({"sm.consolidation.mode": "fragment_meta"})
tiledb.consolidate(uri=path3, config=conf)
fi = tiledb.array_fragments(path3)
self.assertEqual(fi.unconsolidated_metadata_num, 0)
def test_array_consolidate_with_timestamp(self):
dshape = (1, 3)
num_writes = 10
def create_array(target_path, dshape):
dom = tiledb.Domain(tiledb.Dim(domain=dshape, tile=len(dshape)))
att = tiledb.Attr(dtype="int64")
schema = tiledb.ArraySchema(domain=dom, attrs=(att,), sparse=True)
tiledb.libtiledb.Array.create(target_path, schema)
def write_fragments(target_path, dshape, num_writes):
for i in range(1, num_writes + 1):
with tiledb.open(target_path, "w", timestamp=i) as A:
A[[1, 2, 3]] = np.random.rand(dshape[1])
path = self.path("test_array_consolidate_with_timestamp")
create_array(path, dshape)
write_fragments(path, dshape, num_writes)
frags = tiledb.array_fragments(path)
assert len(frags) == 10
tiledb.consolidate(path, timestamp=(1, 4))
frags = tiledb.array_fragments(path)
assert len(frags) == 7
assert len(frags.to_vacuum) == 4
with pytest.warns(
DeprecationWarning,
match=(
"Partial vacuuming via timestamp will be deprecrated in "
"a future release and replaced by passing in fragment URIs."
),
):
tiledb.vacuum(path, timestamp=(1, 2))
tiledb.vacuum(path)
frags = tiledb.array_fragments(path)
assert len(frags.to_vacuum) == 0
conf = tiledb.Config(
{"sm.consolidation.timestamp_start": 5, "sm.consolidation.timestamp_end": 9}
)
tiledb.consolidate(path, config=conf)
tiledb.vacuum(path)
assert len(tiledb.array_fragments(path)) == 3
@pytest.mark.parametrize("use_timestamps", [True, False])
def test_array_consolidate_with_uris(self, use_timestamps):
dshape = (1, 3)
num_writes = 10
def create_array(target_path, dshape):
dom = tiledb.Domain(tiledb.Dim(domain=dshape, tile=len(dshape)))
att = tiledb.Attr(dtype="int64")
schema = tiledb.ArraySchema(domain=dom, attrs=(att,), sparse=True)
tiledb.libtiledb.Array.create(target_path, schema)
def write_fragments(target_path, dshape, num_writes):
for i in range(1, num_writes + 1):
with tiledb.open(
target_path, "w", timestamp=i if use_timestamps else None
) as A:
A[[1, 2, 3]] = np.random.rand(dshape[1])
path = self.path("test_array_consolidate_with_uris")
create_array(path, dshape)
write_fragments(path, dshape, num_writes)
frags = tiledb.array_fragments(path)
assert len(frags) == 10
frag_names = [os.path.basename(f) for f in frags.uri]
tiledb.consolidate(path, fragment_uris=frag_names[:4])
assert len(tiledb.array_fragments(path)) == 7
with pytest.warns(
DeprecationWarning,
match=(
"The `timestamp` argument will be ignored and only fragments "
"passed to `fragment_uris` will be consolidate"
),
):
timestamps = [t[0] for t in tiledb.array_fragments(path).timestamp_range]
tiledb.consolidate(
path,
fragment_uris=frag_names[4:8],
timestamp=(timestamps[5], timestamps[6]),
)
assert len(tiledb.array_fragments(path)) == 4
def test_array_consolidate_with_key(self):
dshape = (1, 3)
num_writes = 10
path = self.path("test_array_consolidate_with_key")
key = "0123456789abcdeF0123456789abcdeF"
config = tiledb.Config()
config["sm.encryption_key"] = key
config["sm.encryption_type"] = "AES_256_GCM"
ctx = tiledb.Ctx(config=config)
def create_array(target_path, dshape):
dom = tiledb.Domain(tiledb.Dim(domain=dshape, tile=len(dshape)))
att = tiledb.Attr(dtype="int64")
schema = tiledb.ArraySchema(domain=dom, attrs=(att,), sparse=True)
tiledb.libtiledb.Array.create(target_path, schema, ctx=ctx)
def write_fragments(target_path, dshape, num_writes):
for i in range(1, num_writes + 1):
with tiledb.open(target_path, "w", timestamp=i, ctx=ctx) as A:
A[[1, 2, 3]] = np.random.rand(dshape[1])
create_array(path, dshape)
write_fragments(path, dshape, num_writes)
frags = tiledb.array_fragments(path, ctx=ctx)
assert len(frags) == 10
frag_names = [os.path.basename(f) for f in frags.uri]
tiledb.consolidate(path, ctx=ctx, config=config, fragment_uris=frag_names[:4])
assert len(tiledb.array_fragments(path, ctx=ctx)) == 7
@pytest.mark.skipif(sys.platform == "win32", reason="Only run MemoryTest on linux")
class MemoryTest(DiskTestCase):
# sanity check that memory usage doesn't increase more than 2x when reading 40MB 100x
# https://github.com/TileDB-Inc/TileDB-Py/issues/150
@staticmethod
def use_many_buffers(path):
# https://stackoverflow.com/questions/938733/total-memory-used-by-python-process
process = psutil.Process(os.getpid())
x = np.ones(10000000, dtype=np.float32)
d1 = tiledb.Dim(
"test_domain", domain=(0, x.shape[0] - 1), tile=10000, dtype="uint32"
)
domain = tiledb.Domain(d1)
v = tiledb.Attr("test_value", dtype="float32")
schema = tiledb.ArraySchema(
domain=domain, attrs=(v,), cell_order="row-major", tile_order="row-major"
)
A = tiledb.DenseArray.create(path, schema)
with tiledb.DenseArray(path, mode="w") as A:
A[:] = {"test_value": x}
with tiledb.DenseArray(path, mode="r") as data:
data[:]
initial = process.memory_info().rss
print(" initial RSS: {}".format(round(initial / 1e6, 2)))
for i in range(100):
# read but don't store: this memory should be freed
data[:]
if i % 10 == 0:
print(
" read iter {}, RSS (MB): {}".format(
i, round(process.memory_info().rss / 1e6, 2)
)
)
return initial
def test_memory_cleanup(self, capfd):
# run function which reads 100x from a 40MB test array
# TODO: RSS is too loose to do this end-to-end, so should use instrumentation.
print("Starting TileDB-Py memory test:")
initial = self.use_many_buffers(self.path("test_memory_cleanup"))
process = psutil.Process(os.getpid())
final = process.memory_info().rss
print(" final RSS: {}".format(round(final / 1e6, 2)))
gc.collect()
final_gc = process.memory_info().rss
print(" final RSS after forced GC: {}".format(round(final_gc / 1e6, 2)))
assert_captured(capfd, "final RSS")
self.assertTrue(final < (2 * initial))
class TestHighlevel(DiskTestCase):
def test_open(self):
uri = self.path("test_open")
array = np.random.rand(10)
schema = tiledb.schema_like(array)
tiledb.Array.create(uri, schema)
with tiledb.open(uri, "w") as A:
A[:] = array * 10
A[:] = array
last_fragment_ts = list(A.last_write_info.items())[0][1][0]
ctx = tiledb.Ctx()
with tiledb.DenseArray(uri, ctx=ctx) as A:
self.assertEqual(A._ctx_(), ctx)
# test `open` with timestamp
with tiledb.open(uri, timestamp=last_fragment_ts) as A:
assert_array_equal(A[:], array)
with tiledb.open(uri, ctx=ctx) as A:
self.assertEqual(A._ctx_(), ctx)
config = tiledb.Config()
with tiledb.open(uri, config=config) as A:
self.assertEqual(A._ctx_().config(), config)
with self.assertRaises(KeyError):
# This path must test `tiledb.open` specifically
# https://github.com/TileDB-Inc/TileDB-Py/issues/277
tiledb.open(uri, "r", attr="the-missing-attr")
def test_ctx_thread_cleanup(self):
# This test checks that contexts are destroyed correctly.
# It creates new contexts repeatedly, in-process, and
# checks that the total number of threads stays stable.
threads = (
"sm.num_reader_threads"
if tiledb.libtiledb.version() < (2, 10)
else "sm.compute_concurrency_level"
)
config = {threads: 128}
uri = self.path("test_ctx_thread_cleanup")
with tiledb.from_numpy(uri, np.random.rand(100)) as A:
pass
thisproc = psutil.Process(os.getpid())
with tiledb.DenseArray(uri, ctx=tiledb.Ctx(config)) as A:
A[:]
start_threads = len(thisproc.threads())
for n in range(1, 10):
retry = 0
while retry < 3:
try:
# checking exact thread count is unreliable, so
# make sure we are holding < 2x per run.
self.assertTrue(len(thisproc.threads()) < 2 * start_threads)
break
except RuntimeError as rterr:
retry += 1
if retry > 2:
raise rterr
warnings.warn(
f"Thread cleanup test RuntimeError: {rterr} \n on iteration: {n}"
)
with tiledb.DenseArray(uri, ctx=tiledb.Ctx(config)) as A:
A[:]
class GetStatsTest(DiskTestCase):
def test_ctx(self):
tiledb.stats_enable()
ctx = tiledb.default_ctx()
uri = self.path("test_ctx")
dom = tiledb.Domain(tiledb.Dim(domain=(0, 2), dtype=np.int64))
att = tiledb.Attr(dtype=np.int64)
schema = tiledb.ArraySchema(domain=dom, attrs=(att,))
tiledb.Array.create(uri, schema)
with tiledb.open(uri, mode="w", ctx=ctx) as T:
T[:] = np.random.randint(10, size=3)
stats = ctx.get_stats(print_out=False)
# check that the stats are non-empty
assert stats
def test_query(self):
tiledb.stats_enable()
uri = self.path("test_ctx")
dom = tiledb.Domain(tiledb.Dim(domain=(0, 2), dtype=np.int64))
att = tiledb.Attr(dtype=np.int64)
schema = tiledb.ArraySchema(domain=dom, attrs=(att,))
tiledb.Array.create(uri, schema)
with tiledb.open(uri, mode="w") as T:
T[:] = np.random.randint(10, size=3)
with tiledb.open(uri, mode="r") as T:
q = T.query()
assert "" == q.get_stats()
q[:]
stats = q.get_stats(print_out=False)
# check that the stats are non-empty
assert stats
class NullableIOTest(DiskTestCase):
def test_nullable_write(self):
uri = self.path("nullable_write_test")
schema = tiledb.ArraySchema(
domain=tiledb.Domain(
*[tiledb.Dim(name="__dim_0", domain=(0, 3), tile=4, dtype="uint64")]
),
attrs=[tiledb.Attr(name="", dtype="int64", var=False, nullable=True)],
)
tiledb.Array.create(uri, schema)
with tiledb.open(uri, "w") as A:
A._setitem_impl(
slice(0, 4), np.ones(4), {"": np.array([0, 1, 0, 1], dtype=np.uint8)}
)
class IncompleteTest(DiskTestCase):
@pytest.mark.parametrize("non_overlapping_ranges", [True, False])
def test_incomplete_dense_varlen(self, non_overlapping_ranges):
ncells = 10
path = self.path("incomplete_dense_varlen")
str_data = [rand_utf8(random.randint(0, n)) for n in range(ncells)]
data = np.array(str_data, dtype=np.str_)
# basic write
dom = tiledb.Domain(tiledb.Dim(domain=(1, len(data)), tile=len(data)))
att = tiledb.Attr(dtype=np.str_, var=True)
schema = tiledb.ArraySchema(dom, (att,))
tiledb.DenseArray.create(path, schema)
with tiledb.DenseArray(path, mode="w") as T:
T[:] = data
with tiledb.DenseArray(path, mode="r") as T:
assert_array_equal(data, T[:])
# set the memory to the max length of a cell
# these settings force ~100 retries
# TODO would be good to check repeat count here; not yet exposed
# Also would be useful to have max cell config in libtiledb.
init_buffer_bytes = 1024**2
config = tiledb.Config(
{
"sm.memory_budget": ncells,
"sm.memory_budget_var": ncells,
"py.init_buffer_bytes": init_buffer_bytes,
"sm.query.sparse_unordered_with_dups.non_overlapping_ranges": non_overlapping_ranges,
"sm.skip_unary_partitioning_budget_check": True,
}
)
self.assertEqual(config["py.init_buffer_bytes"], str(init_buffer_bytes))
with tiledb.DenseArray(path, mode="r", ctx=tiledb.Ctx(config)) as T2:
result = T2.query(attrs=[""])[:]
assert_array_equal(result, data)
@pytest.mark.parametrize("allows_duplicates", [True, False])
@pytest.mark.parametrize("non_overlapping_ranges", [True, False])
def test_incomplete_sparse_varlen(self, allows_duplicates, non_overlapping_ranges):
ncells = 100
path = self.path("incomplete_sparse_varlen")
str_data = [rand_utf8(random.randint(0, n)) for n in range(ncells)]
data = np.array(str_data, dtype=np.str_)
coords = np.arange(ncells)
# basic write
dom = tiledb.Domain(tiledb.Dim(domain=(0, len(data) + 100), tile=len(data)))
att = tiledb.Attr(dtype=np.str_, var=True)
schema = tiledb.ArraySchema(
dom, (att,), sparse=True, allows_duplicates=allows_duplicates
)
tiledb.SparseArray.create(path, schema)
with tiledb.SparseArray(path, mode="w") as T:
T[coords] = data
with tiledb.SparseArray(path, mode="r") as T:
assert_array_equal(data, T[:][""])
# set the memory to the max length of a cell
# these settings force ~100 retries
# TODO would be good to check repeat count here; not yet exposed
# Also would be useful to have max cell config in libtiledb.
init_buffer_bytes = 1024**2
config = tiledb.Config(
{
"sm.memory_budget": ncells,
"sm.memory_budget_var": ncells,
"py.init_buffer_bytes": init_buffer_bytes,
}
)
self.assertEqual(config["py.init_buffer_bytes"], str(init_buffer_bytes))
with tiledb.SparseArray(path, mode="r", ctx=tiledb.Ctx(config)) as T2:
assert_array_equal(data, T2[:][""])
assert_array_equal(data, T2.multi_index[0:ncells][""])
# ensure that empty results are handled correctly
assert_array_equal(
T2.multi_index[101:105][""], np.array([], dtype=np.dtype("<U"))
)
@pytest.mark.parametrize("sparse", [True, False])
def test_query_return_incomplete_error(self, sparse):
path = self.path("test_query_return_incomplete_error")
dom = tiledb.Domain(tiledb.Dim(domain=(1, 3), tile=1))
attrs = [tiledb.Attr("ints", dtype=np.uint8)]
schema = tiledb.ArraySchema(domain=dom, attrs=attrs, sparse=sparse)
tiledb.Array.create(path, schema)
with tiledb.open(path, "r") as A:
if sparse:
A.query(return_incomplete=True)[:]
else:
with self.assertRaises(tiledb.TileDBError):
A.query(return_incomplete=True)[:]
@pytest.mark.skipif(not has_pandas(), reason="pandas>=1.0,<3.0 not installed")
@pytest.mark.parametrize(
"use_arrow, return_arrow, indexer",
[
(True, True, "df"),
(True, False, "df"),
(False, False, "df"),
(None, False, "multi_index"),
],
)
@pytest.mark.parametrize(
"test_incomplete_return_array", [True, False], indirect=True
)
@pytest.mark.parametrize("non_overlapping_ranges", [True, False])
def test_incomplete_return(
self,
test_incomplete_return_array,
use_arrow,
return_arrow,
indexer,
non_overlapping_ranges,
):
import pandas as pd
import pyarrow as pa
from tiledb.multirange_indexing import EstimatedResultSize
path = test_incomplete_return_array
init_buffer_bytes = 200
cfg = tiledb.Config(
{
"py.init_buffer_bytes": init_buffer_bytes,
"py.exact_init_buffer_bytes": "true",
"sm.query.sparse_unordered_with_dups.non_overlapping_ranges": non_overlapping_ranges,
}
)
with tiledb.open(path) as A:
full_data = A[:][""]
# count number of elements retrieved so that we can slice the comparison array
idx = 0
with tiledb.open(path, ctx=tiledb.Ctx(cfg)) as A:
query = A.query(
return_incomplete=True, use_arrow=use_arrow, return_arrow=return_arrow
)
# empty range
iterable = getattr(query, indexer)[tiledb.EmptyRange]
est_results = iterable.estimated_result_sizes()
assert isinstance(est_results[""], EstimatedResultSize)
assert isinstance(est_results["__dim_0"], EstimatedResultSize)
assert est_results["__dim_0"].offsets_bytes == 0
assert est_results["__dim_0"].data_bytes == 0
assert est_results[""].offsets_bytes == 0
assert est_results[""].data_bytes == 0
results = list(iterable)
assert len(results) == 1
result = results[0]
if indexer == "df":
assert isinstance(result, pd.DataFrame)
assert len(result) == 0
else:
assert isinstance(result, OrderedDict)
assert len(pd.DataFrame(result)) == 0
# full range
iterable = getattr(query, indexer)[:]
est_results = iterable.estimated_result_sizes()
assert isinstance(est_results[""], EstimatedResultSize)
assert isinstance(est_results["__dim_0"], EstimatedResultSize)
assert est_results["__dim_0"].offsets_bytes == 0
assert est_results["__dim_0"].data_bytes > 0
assert est_results[""].offsets_bytes > 0
assert est_results[""].data_bytes > 0
for result in iterable:
if return_arrow:
assert isinstance(result, pa.Table)
df = result.to_pandas()
else:
if indexer == "df":
assert isinstance(result, pd.DataFrame)
df = result
else:
assert isinstance(result, OrderedDict)
df = pd.DataFrame(result)
to_slice = slice(idx, idx + len(df))
chunk = full_data[to_slice]
assert np.all(chunk == df[""].values)
assert np.all(df["__dim_0"] == np.arange(idx, idx + len(df)))
# update the current read count
idx += len(df)
assert idx == len(full_data)
@pytest.mark.parametrize("cell_order", ["col-major", "row-major", "hilbert"])
@pytest.mark.parametrize("tile_order", ["col-major", "row-major"])
@pytest.mark.parametrize("non_overlapping_ranges", [True, False])
def test_incomplete_global_order(
self, cell_order, tile_order, non_overlapping_ranges
):
uri = self.path("test_incomplete_global_order")
dom = tiledb.Domain(tiledb.Dim(domain=(0, 30), tile=10, dtype=np.int64))
att = tiledb.Attr(dtype=np.int64)
schema = tiledb.ArraySchema(
domain=dom,
attrs=(att,),
sparse=True,
allows_duplicates=True,
cell_order=cell_order,
tile_order=tile_order,
)
tiledb.Array.create(uri, schema)
expected_data = np.random.randint(0, 10, 30)
with tiledb.open(uri, mode="w") as T:
T[np.arange(30)] = expected_data
init_buffer_bytes = 200
cfg = tiledb.Config(
{
"py.init_buffer_bytes": init_buffer_bytes,
"py.exact_init_buffer_bytes": "true",
"sm.query.sparse_unordered_with_dups.non_overlapping_ranges": non_overlapping_ranges,
}
)
with tiledb.open(uri, mode="r", ctx=tiledb.Ctx(cfg)) as T:
actual_data = T.query(order="G")[:][""]
assert_array_equal(actual_data, expected_data)
@pytest.mark.parametrize("exact_init_buffer_bytes", ["true", "false"])
@pytest.mark.parametrize("non_overlapping_ranges", [True, False])
def test_offset_can_fit_data_var_size_cannot(
self, exact_init_buffer_bytes, non_overlapping_ranges
):
"""
One condition that would be nice to get more coverage on is when the offset buffer can fit X cells, but the var size data of those cells cannot fit the buffer. In this case, the reader does adjust the results back.
@Luc Rancourt so would we test this by having really large var-size content in each cell?
Isaiah 4 days ago
eg something like: we set buffers that can hold 100kb, but each var-len cell has 20kb, so we can read at most 5 cells into the data buffer, but theoretically the offsets buffer could hold many more?
"""
tiledb.stats_enable()
uri = self.path("test_incomplete_global_order")
dom = tiledb.Domain(tiledb.Dim(domain=(0, 4), tile=1, dtype=np.int64))
att = tiledb.Attr(dtype=np.int64, var=True)
schema = tiledb.ArraySchema(
domain=dom,
attrs=(att,),
sparse=True,
allows_duplicates=True,
)
tiledb.Array.create(uri, schema)
with tiledb.open(uri, mode="w") as T:
T[np.arange(5)] = np.array(
[
np.random.randint(0, 10, 10000, dtype=np.int64),
np.random.randint(0, 10, 10000, dtype=np.int64),
np.random.randint(0, 10, 10000, dtype=np.int64),
np.random.randint(0, 10, 10000, dtype=np.int64),
np.random.randint(0, 10, 101, dtype=np.int64),
],
dtype="O",
)
init_buffer_bytes = 160000
cfg = tiledb.Config(
{
"py.init_buffer_bytes": init_buffer_bytes,
"py.exact_init_buffer_bytes": exact_init_buffer_bytes,
"sm.query.sparse_unordered_with_dups.non_overlapping_ranges": non_overlapping_ranges,
}
)
with tiledb.open(uri, mode="r", ctx=tiledb.Ctx(cfg)) as T:
qry = T.query()
qry[:][""]
# assert_array_equal(actual_data, expected_data)
tiledb.stats_disable()
class TestPath(DiskTestCase):
def test_path(self, pytestconfig):
path = self.path("foo")
if pytestconfig.getoption("vfs") == "s3":
assert path.startswith("s3://")
@pytest.mark.skipif(
sys.platform == "win32", reason="no_output fixture disabled on Windows"
)
@pytest.mark.xfail(
True, reason="This test prints, and should fail because of no_output fixture!"
)
def test_no_output(self):
print("this test should fail")
class TestAsBuilt(DiskTestCase):
def test_as_built(self):
dump = tiledb.as_built(return_json_string=True)
assert isinstance(dump, str)
# ensure we get a non-empty string
assert len(dump) > 0
dump_dict = tiledb.as_built()
assert isinstance(dump_dict, dict)
# ensure we get a non-empty dict
assert len(dump_dict) > 0
# validate top-level key
assert "as_built" in dump_dict
assert isinstance(dump_dict["as_built"], dict)
assert len(dump_dict["as_built"]) > 0
# validate parameters key
assert "parameters" in dump_dict["as_built"]
assert isinstance(dump_dict["as_built"]["parameters"], dict)
assert len(dump_dict["as_built"]["parameters"]) > 0
# validate storage_backends key
assert "storage_backends" in dump_dict["as_built"]["parameters"]
assert isinstance(dump_dict["as_built"]["parameters"]["storage_backends"], dict)
assert len(dump_dict["as_built"]["parameters"]["storage_backends"]) > 0
x = dump_dict["as_built"]["parameters"]["storage_backends"]
# validate storage_backends attributes
vfs = tiledb.VFS()
if vfs.supports("azure"):
assert x["azure"]["enabled"] == True
else:
assert x["azure"]["enabled"] == False
if vfs.supports("gcs"):
assert x["gcs"]["enabled"] == True
else:
assert x["gcs"]["enabled"] == False
if vfs.supports("hdfs"):
assert x["hdfs"]["enabled"] == True
else:
assert x["hdfs"]["enabled"] == False
if vfs.supports("s3"):
assert x["s3"]["enabled"] == True
else:
assert x["s3"]["enabled"] == False
# validate support key
assert "support" in dump_dict["as_built"]["parameters"]
assert isinstance(dump_dict["as_built"]["parameters"]["support"], dict)
assert len(dump_dict["as_built"]["parameters"]["support"]) > 0
# validate support attributes - check only if boolean
assert dump_dict["as_built"]["parameters"]["support"]["serialization"][
"enabled"
] in [True, False]