| | import collections |
| | import datetime |
| | import io |
| | import pathlib |
| | import uuid |
| | import random |
| | from sklearn.model_selection import train_test_split |
| |
|
| | import numpy as np |
| | import tensorflow as tf |
| |
|
| |
|
| | class Replay: |
| | def __init__( |
| | self, directory, capacity=0, ongoing=False, minlen=1, maxlen=0, |
| | prioritize_ends=False, disable_save=False): |
| | self._directory = pathlib.Path(directory).expanduser() |
| | self._directory.mkdir(parents=True, exist_ok=True) |
| | self._capacity = capacity |
| | self._ongoing = ongoing |
| | self._minlen = minlen |
| | self._maxlen = maxlen |
| | self._prioritize_ends = prioritize_ends |
| | self._random = np.random.RandomState() |
| | |
| | self._complete_eps = load_episodes(self._directory, capacity, minlen) |
| | |
| | self._ongoing_eps = collections.defaultdict( |
| | lambda: collections.defaultdict(list)) |
| | self._total_episodes, self._total_steps = count_episodes(directory) |
| | self._loaded_episodes = len(self._complete_eps) |
| | self._loaded_steps = sum(eplen(x) for x in self._complete_eps.values()) |
| | self._disable_save = disable_save |
| |
|
| | @property |
| | def stats(self): |
| | return { |
| | 'total_steps': self._total_steps, |
| | 'total_episodes': self._total_episodes, |
| | 'loaded_steps': self._loaded_steps, |
| | 'loaded_episodes': self._loaded_episodes, |
| | } |
| |
|
| | def add_step(self, transition, worker=0): |
| | episode = self._ongoing_eps[worker] |
| | for key, value in transition.items(): |
| | episode[key].append(value) |
| | if transition['is_last']: |
| | self.add_episode(episode) |
| | episode.clear() |
| |
|
| | def add_episode(self, episode): |
| | length = eplen(episode) |
| | if length < self._minlen: |
| | print(f'Skipping short episode of length {length}.') |
| | return |
| | self._total_steps += length |
| | self._loaded_steps += length |
| | self._total_episodes += 1 |
| | self._loaded_episodes += 1 |
| | episode = {key: convert(value) for key, value in episode.items()} |
| | filename = save_episode(self._directory, episode, disable=self._disable_save) |
| | self._complete_eps[str(filename)] = episode |
| | self._enforce_limit() |
| |
|
| | def dataset(self, batch, length): |
| | example = next(iter(self._generate_chunks(length))) |
| | dataset = tf.data.Dataset.from_generator( |
| | lambda: self._generate_chunks(length), |
| | {k: v.dtype for k, v in example.items()}, |
| | {k: v.shape for k, v in example.items()}) |
| | dataset = dataset.batch(batch, drop_remainder=True) |
| | dataset = dataset.prefetch(5) |
| | return dataset |
| |
|
| | def _generate_chunks(self, length): |
| | sequence = self._sample_sequence() |
| | while True: |
| | chunk = collections.defaultdict(list) |
| | added = 0 |
| | while added < length: |
| | needed = length - added |
| | adding = {k: v[:needed] for k, v in sequence.items()} |
| | sequence = {k: v[needed:] for k, v in sequence.items()} |
| | for key, value in adding.items(): |
| | chunk[key].append(value) |
| | added += len(adding['action']) |
| | if len(sequence['action']) < 1: |
| | sequence = self._sample_sequence() |
| | chunk = {k: np.concatenate(v) for k, v in chunk.items()} |
| | yield chunk |
| |
|
| | def _sample_sequence(self): |
| | episodes = list(self._complete_eps.values()) |
| | if self._ongoing: |
| | episodes += [ |
| | x for x in self._ongoing_eps.values() |
| | if eplen(x) >= self._minlen] |
| | episode = self._random.choice(episodes) |
| | total = len(episode['action']) |
| | length = total |
| | if self._maxlen: |
| | length = min(length, self._maxlen) |
| | |
| | |
| | length -= np.random.randint(self._minlen) |
| | length = max(self._minlen, length) |
| | upper = total - length + 1 |
| | if self._prioritize_ends: |
| | upper += self._minlen |
| | index = min(self._random.randint(upper), total - length) |
| | sequence = { |
| | k: convert(v[index: index + length]) |
| | for k, v in episode.items() if not k.startswith('log_')} |
| | sequence['is_first'] = np.zeros(len(sequence['action']), np.bool) |
| | sequence['is_first'][0] = True |
| | if self._maxlen: |
| | assert self._minlen <= len(sequence['action']) <= self._maxlen |
| | return sequence |
| |
|
| | def _enforce_limit(self): |
| | if not self._capacity: |
| | return |
| | while self._loaded_episodes > 1 and self._loaded_steps > self._capacity: |
| | |
| | oldest, episode = next(iter(self._complete_eps.items())) |
| | self._loaded_steps -= eplen(episode) |
| | self._loaded_episodes -= 1 |
| | del self._complete_eps[oldest] |
| |
|
| |
|
| | def count_episodes(directory): |
| | filenames = list(directory.glob('*.npz')) |
| | num_episodes = len(filenames) |
| | num_steps = sum(int(str(n).split('-')[-1][:-4]) - 1 for n in filenames) |
| | return num_episodes, num_steps |
| |
|
| |
|
| | def save_episode(directory, episode, disable=False): |
| | timestamp = datetime.datetime.now().strftime('%Y%m%dT%H%M%S') |
| | identifier = str(uuid.uuid4().hex) |
| | length = eplen(episode) |
| | filename = directory / f'{timestamp}-{identifier}-{length}.npz' |
| | if not disable: |
| | with io.BytesIO() as f1: |
| | np.savez_compressed(f1, **episode) |
| | f1.seek(0) |
| | with filename.open('wb') as f2: |
| | f2.write(f1.read()) |
| | return filename |
| |
|
| |
|
| | def load_episodes(directory, capacity=None, minlen=1, keep_temporal_order=False): |
| | |
| | |
| | filenames = sorted(directory.glob('*.npz')) |
| | if not keep_temporal_order: |
| | print('Shuffling order of offline trajectories!') |
| | random.Random(0).shuffle(filenames) |
| | if capacity: |
| | num_steps = 0 |
| | num_episodes = 0 |
| | for filename in reversed(filenames): |
| | length = int(str(filename).split('-')[-1][:-4]) |
| | num_steps += length |
| | num_episodes += 1 |
| | if num_steps >= capacity: |
| | break |
| | filenames = filenames[-num_episodes:] |
| | episodes = {} |
| | for filename in filenames: |
| | try: |
| | with filename.open('rb') as f: |
| | episode = np.load(f) |
| | episode = {k: episode[k] for k in episode.keys()} |
| | |
| | if 'is_terminal' not in episode: |
| | episode['is_terminal'] = episode['discount'] == 0. |
| | except Exception as e: |
| | print(f'Could not load episode {str(filename)}: {e}') |
| | continue |
| | episodes[str(filename)] = episode |
| | return episodes |
| |
|
| |
|
| | def convert(value): |
| | value = np.array(value) |
| | if np.issubdtype(value.dtype, np.floating): |
| | return value.astype(np.float32) |
| | elif np.issubdtype(value.dtype, np.signedinteger): |
| | return value.astype(np.int32) |
| | elif np.issubdtype(value.dtype, np.uint8): |
| | return value.astype(np.uint8) |
| | return value |
| |
|
| |
|
| | def eplen(episode): |
| | return len(episode['action']) - 1 |
| |
|
| |
|
| | class OfflineReplay: |
| | def __init__( |
| | self, directory, capacity=0, ongoing=False, minlen=1, maxlen=0, |
| | prioritize_ends=False, split_val=False): |
| | self._capacity = capacity |
| | self._minlen = minlen |
| | self._maxlen = maxlen |
| | self._prioritize_ends = prioritize_ends |
| | self._random = np.random.RandomState() |
| | self._total_episodes = 0 |
| | self._total_steps = 0 |
| | self._loaded_episodes = 0 |
| | self._loaded_steps = 0 |
| | self._complete_eps = {} |
| |
|
| | if type(directory) is not list: |
| | directory = [directory] |
| |
|
| | for d in directory: |
| | path = pathlib.Path(d).expanduser() |
| | complete_eps = load_episodes(path, capacity, minlen) |
| | self._complete_eps.update(complete_eps) |
| | t_eps, t_steps = count_episodes(path) |
| | self._total_episodes += t_eps |
| | self._total_steps += t_steps |
| | self._loaded_episodes += len(complete_eps) |
| | self._loaded_steps += sum(eplen(x) for x in complete_eps.values()) |
| |
|
| | print('Loaded from offline directory(ies)!') |
| | print(self.stats) |
| |
|
| | self.episodes = list(self._complete_eps.values()) |
| |
|
| | if split_val: |
| | self.episodes, self.val_episodes = train_test_split(self.episodes, test_size=0.2) |
| | print( |
| | f'Split into {len(self.episodes)} training episodes and {len(self.val_episodes)} validation episodes.') |
| |
|
| | @property |
| | def stats(self): |
| | return { |
| | 'total_steps': self._total_steps, |
| | 'total_episodes': self._total_episodes, |
| | 'loaded_steps': self._loaded_steps, |
| | 'loaded_episodes': self._loaded_episodes, |
| | } |
| |
|
| | def dataset(self, batch, length, validation=False): |
| | example = next(iter(self._generate_chunks(length, validation=validation))) |
| | dataset = tf.data.Dataset.from_generator( |
| | lambda: self._generate_chunks(length, validation=validation), |
| | {k: v.dtype for k, v in example.items()}, |
| | {k: v.shape for k, v in example.items()}) |
| | dataset = dataset.batch(batch, drop_remainder=True) |
| | dataset = dataset.prefetch(5) |
| | return dataset |
| |
|
| | def _generate_chunks(self, length, validation): |
| | sequence = self._sample_sequence(validation=validation) |
| | while True: |
| | chunk = collections.defaultdict(list) |
| | added = 0 |
| | while added < length: |
| | needed = length - added |
| | adding = {k: v[:needed] for k, v in sequence.items()} |
| | sequence = {k: v[needed:] for k, v in sequence.items()} |
| | for key, value in adding.items(): |
| | chunk[key].append(value) |
| | added += len(adding['action']) |
| | if len(sequence['action']) < 1: |
| | sequence = self._sample_sequence(validation=validation) |
| | chunk = {k: np.concatenate(v) for k, v in chunk.items()} |
| | yield chunk |
| |
|
| | def _sample_sequence(self, validation): |
| | if validation: |
| | episode = self._random.choice(self.val_episodes) |
| | else: |
| | episode = self._random.choice(self.episodes) |
| | total = len(episode['action']) |
| | length = total |
| | if self._maxlen: |
| | length = min(length, self._maxlen) |
| | |
| | |
| | length -= np.random.randint(self._minlen) |
| | length = max(self._minlen, length) |
| | upper = total - length + 1 |
| | if self._prioritize_ends: |
| | upper += self._minlen |
| | index = min(self._random.randint(upper), total - length) |
| | sequence = { |
| | k: convert(v[index: index + length]) |
| | for k, v in episode.items() if not k.startswith('log_')} |
| | sequence['is_first'] = np.zeros(len(sequence['action']), np.bool) |
| | sequence['is_first'][0] = True |
| | if self._maxlen: |
| | assert self._minlen <= len(sequence['action']) <= self._maxlen |
| | return sequence |
| |
|
| | |
| | def validation_chunks(self, batch, length): |
| | chunks = [] |
| | for episode in self.val_episodes: |
| | episode_chunks = [] |
| | sequence = {k: convert(v) for k, v in episode.items() if not k.startswith('log_')} |
| | sequence['is_first'] = np.zeros(len(sequence['action']), np.bool) |
| |
|
| | while len(sequence['action']) >= length: |
| | chunk = collections.defaultdict(list) |
| | adding = {k: v[:length] for k, v in sequence.items()} |
| | adding['is_first'][0] = True |
| | sequence = {k: v[length:] for k, v in sequence.items()} |
| | for key, value in adding.items(): |
| | chunk[key].append(value) |
| | chunk = {k: np.concatenate(v) for k, v in chunk.items()} |
| | episode_chunks.append(chunk) |
| | episode_chunks = {k: tf.stack([x[k] for x in episode_chunks]) for k in episode_chunks[0].keys()} |
| | chunks.append(episode_chunks) |
| |
|
| | return chunks |
| |
|