|
|
import functools |
|
|
import queue |
|
|
import threading |
|
|
|
|
|
import elements |
|
|
import numpy as np |
|
|
import portal |
|
|
|
|
|
from . import base |
|
|
|
|
|
|
|
|
class Stateless(base.Stream): |
|
|
|
|
|
def __init__(self, nextfn, *args, **kwargs): |
|
|
if not callable(nextfn) and hasattr(nextfn, '__next__'): |
|
|
nextfn = nextfn.__next__ |
|
|
self.nextfn = functools.partial(nextfn, *args, **kwargs) |
|
|
|
|
|
def __iter__(self): |
|
|
return self |
|
|
|
|
|
def __next__(self): |
|
|
return self.nextfn() |
|
|
|
|
|
def save(self): |
|
|
return None |
|
|
|
|
|
def load(self, data): |
|
|
pass |
|
|
|
|
|
|
|
|
class Prefetch(base.Stream): |
|
|
|
|
|
def __init__(self, source, transform=None, amount=1): |
|
|
self.source = iter(source) if hasattr(source, '__iter__') else source() |
|
|
self.transform = transform or (lambda x: x) |
|
|
self.state = self._getstate() |
|
|
self.requests = threading.Semaphore(amount) |
|
|
self.amount = amount |
|
|
self.queue = queue.Queue() |
|
|
self.worker = portal.Thread(self._worker) |
|
|
self.started = False |
|
|
|
|
|
def __iter__(self): |
|
|
assert not self.started |
|
|
self.worker.start() |
|
|
self.started = True |
|
|
return self |
|
|
|
|
|
def __next__(self): |
|
|
assert self.started |
|
|
result = self.queue.get() |
|
|
self.requests.release() |
|
|
if isinstance(result, str): |
|
|
raise RuntimeError(result) |
|
|
data, self.state = result |
|
|
return data |
|
|
|
|
|
def save(self): |
|
|
return self.state |
|
|
|
|
|
def load(self, state): |
|
|
if self.started: |
|
|
for _ in range(self.amount): |
|
|
self.queue.get() |
|
|
self.source.load(state) |
|
|
if self.started: |
|
|
self.requests.release(self.amount) |
|
|
|
|
|
def _worker(self): |
|
|
try: |
|
|
while True: |
|
|
self.requests.acquire() |
|
|
data = next(self.source) |
|
|
data = self.transform(data) |
|
|
state = self._getstate() |
|
|
self.queue.put((data, state)) |
|
|
except Exception as e: |
|
|
self.queue.put(str(e)) |
|
|
raise |
|
|
|
|
|
def _getstate(self): |
|
|
if hasattr(self.source, 'save'): |
|
|
return self.source.save() |
|
|
else: |
|
|
return None |
|
|
|
|
|
|
|
|
class Consec(base.Stream): |
|
|
|
|
|
""" |
|
|
Example: |
|
|
|
|
|
length = 3 |
|
|
consec = 3 |
|
|
prefix = 2 |
|
|
|
|
|
source: 0 1 2 3 4 5 6 7 8 9 10 |
|
|
chunk 1: p-p-#-#-# |
|
|
chunk 2: p-p-#-#-# |
|
|
chunk 3: p-p-#-#-# |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, source, length, consec, prefix=0, strict=True, contiguous=False): |
|
|
self.source = source |
|
|
self.length = length |
|
|
self.consec = consec |
|
|
self.prefix = prefix |
|
|
self.strict = strict |
|
|
self.contiguous = contiguous |
|
|
self.index = 0 |
|
|
self.current = None |
|
|
self.it = None |
|
|
|
|
|
def __iter__(self): |
|
|
self.it = iter(self.source) |
|
|
return self |
|
|
|
|
|
def __next__(self): |
|
|
if self.index >= self.consec: |
|
|
self.index = 0 |
|
|
if self.index == 0: |
|
|
self.current = next(self.it) |
|
|
available = self.current['is_first'].shape[-1] |
|
|
assert self.length * self.consec + self.prefix <= available, ( |
|
|
self.length, self.consec, self.prefix, available) |
|
|
if self.strict: |
|
|
assert self.consec * self.length + self.prefix == available, ( |
|
|
self.consec, self.length, self.prefix, available) |
|
|
start = self.index * self.length |
|
|
stop = start + (self.length + self.prefix) |
|
|
chunk = {k: v[:, start: stop] for k, v in self.current.items()} |
|
|
chunk['consec'] = np.full(chunk['is_first'].shape, self.index, np.int32) |
|
|
if self.contiguous: |
|
|
|
|
|
|
|
|
chunk = {k: np.ascontiguousarray(v) for k, v in chunk.items()} |
|
|
self.index += 1 |
|
|
return chunk |
|
|
|
|
|
def save(self): |
|
|
return { |
|
|
'source': self.source.save(), |
|
|
'index': self.index, |
|
|
} |
|
|
|
|
|
def load(self, data): |
|
|
self.source.load(data['source']) |
|
|
self.index = data['index'] |
|
|
|
|
|
|
|
|
class Zip(base.Stream): |
|
|
|
|
|
def __init__(self, sources): |
|
|
assert len(sources) > 1, len(sources) |
|
|
self.sources = sources |
|
|
self.iterators = None |
|
|
self.started = False |
|
|
|
|
|
def __iter__(self): |
|
|
assert not self.started |
|
|
self.started = True |
|
|
self.iterators = [iter(x) for x in self.sources] |
|
|
return self |
|
|
|
|
|
def __next__(self): |
|
|
parts = [next(x) for x in self.iterators] |
|
|
result = elements.tree.map(lambda *el: np.concatenate(el), *parts) |
|
|
return result |
|
|
|
|
|
def save(self): |
|
|
return [x.save() for x in self.iterators] |
|
|
|
|
|
def load(self, data): |
|
|
assert len(data) == len(self.iterators) |
|
|
[it.load(d) for it, d in zip(self.iterators, data)] |
|
|
|
|
|
|
|
|
class Map(base.Stream): |
|
|
|
|
|
def __init__(self, source, fn, *args, **kwargs): |
|
|
self.source = source |
|
|
self.fn = lambda x: fn(x, *args, **kwargs) |
|
|
self.iterator = None |
|
|
self.started = False |
|
|
|
|
|
def __iter__(self): |
|
|
assert not self.started |
|
|
self.started = True |
|
|
self.iterator = iter(self.source) |
|
|
return self |
|
|
|
|
|
def __next__(self): |
|
|
assert self.started |
|
|
return self.fn(next(self.iterator)) |
|
|
|
|
|
def save(self): |
|
|
return self.iterator.save() |
|
|
|
|
|
def load(self, data): |
|
|
self.iterator.load(data) |
|
|
|
|
|
|
|
|
class Mixer(base.Stream): |
|
|
|
|
|
def __init__(self, sources, weights, seed=0): |
|
|
assert sources.keys() == weights.keys(), (sources, weights) |
|
|
self.keys = sorted(sources.keys()) |
|
|
self.iterators = [iter(sources[k]) for k in self.keys] |
|
|
weights = np.array([weights[k] for k in self.keys], np.float32) |
|
|
self.probs = weights / weights.sum() |
|
|
self.seed = seed |
|
|
self.started = False |
|
|
self.step = 0 |
|
|
|
|
|
def __iter__(self): |
|
|
assert not self.started |
|
|
return self |
|
|
|
|
|
def __next__(self): |
|
|
assert self.started |
|
|
rng = np.ranodm.default_rng(seed=[self.seed, self.step]) |
|
|
self.step += 1 |
|
|
index = rng.choice(len(self.keys), p=self.probs) |
|
|
return next(self.iterators[index]) |
|
|
|
|
|
def save(self): |
|
|
return { |
|
|
'step': self.step, |
|
|
'seed': self.seed, |
|
|
'sources': {k: it.save() for k, it in zip(self.keys, self.iterators)}, |
|
|
} |
|
|
|
|
|
def load(self, data): |
|
|
self.step = data['step'] |
|
|
self.seed = data['seed'] |
|
|
assert sorted(data['sources'].keys()) == self.keys, ( |
|
|
data['sources'], self.keys) |
|
|
for key in self.keys: |
|
|
self.iterators[key].load(data['sources'][key]) |
|
|
|