import collections import datetime import io import pathlib import uuid import random from sklearn.model_selection import train_test_split import numpy as np import tensorflow as tf class Replay: def __init__( self, directory, capacity=0, ongoing=False, minlen=1, maxlen=0, prioritize_ends=False, disable_save=False): self._directory = pathlib.Path(directory).expanduser() self._directory.mkdir(parents=True, exist_ok=True) self._capacity = capacity self._ongoing = ongoing self._minlen = minlen self._maxlen = maxlen self._prioritize_ends = prioritize_ends self._random = np.random.RandomState() # filename -> key -> value_sequence self._complete_eps = load_episodes(self._directory, capacity, minlen) # worker -> key -> value_sequence self._ongoing_eps = collections.defaultdict( lambda: collections.defaultdict(list)) self._total_episodes, self._total_steps = count_episodes(directory) self._loaded_episodes = len(self._complete_eps) self._loaded_steps = sum(eplen(x) for x in self._complete_eps.values()) self._disable_save = disable_save @property def stats(self): return { 'total_steps': self._total_steps, 'total_episodes': self._total_episodes, 'loaded_steps': self._loaded_steps, 'loaded_episodes': self._loaded_episodes, } def add_step(self, transition, worker=0): episode = self._ongoing_eps[worker] for key, value in transition.items(): episode[key].append(value) if transition['is_last']: self.add_episode(episode) episode.clear() def add_episode(self, episode): length = eplen(episode) if length < self._minlen: print(f'Skipping short episode of length {length}.') return self._total_steps += length self._loaded_steps += length self._total_episodes += 1 self._loaded_episodes += 1 episode = {key: convert(value) for key, value in episode.items()} filename = save_episode(self._directory, episode, disable=self._disable_save) self._complete_eps[str(filename)] = episode self._enforce_limit() def dataset(self, batch, length): example = next(iter(self._generate_chunks(length))) dataset = tf.data.Dataset.from_generator( lambda: self._generate_chunks(length), {k: v.dtype for k, v in example.items()}, {k: v.shape for k, v in example.items()}) dataset = dataset.batch(batch, drop_remainder=True) dataset = dataset.prefetch(5) return dataset def _generate_chunks(self, length): sequence = self._sample_sequence() while True: chunk = collections.defaultdict(list) added = 0 while added < length: needed = length - added adding = {k: v[:needed] for k, v in sequence.items()} sequence = {k: v[needed:] for k, v in sequence.items()} for key, value in adding.items(): chunk[key].append(value) added += len(adding['action']) if len(sequence['action']) < 1: sequence = self._sample_sequence() chunk = {k: np.concatenate(v) for k, v in chunk.items()} yield chunk def _sample_sequence(self): episodes = list(self._complete_eps.values()) if self._ongoing: episodes += [ x for x in self._ongoing_eps.values() if eplen(x) >= self._minlen] episode = self._random.choice(episodes) total = len(episode['action']) length = total if self._maxlen: length = min(length, self._maxlen) # Randomize length to avoid all chunks ending at the same time in case the # episodes are all of the same length. length -= np.random.randint(self._minlen) length = max(self._minlen, length) upper = total - length + 1 if self._prioritize_ends: upper += self._minlen index = min(self._random.randint(upper), total - length) sequence = { k: convert(v[index: index + length]) for k, v in episode.items() if not k.startswith('log_')} sequence['is_first'] = np.zeros(len(sequence['action']), np.bool) sequence['is_first'][0] = True if self._maxlen: assert self._minlen <= len(sequence['action']) <= self._maxlen return sequence def _enforce_limit(self): if not self._capacity: return while self._loaded_episodes > 1 and self._loaded_steps > self._capacity: # Relying on Python preserving the insertion order of dicts. oldest, episode = next(iter(self._complete_eps.items())) self._loaded_steps -= eplen(episode) self._loaded_episodes -= 1 del self._complete_eps[oldest] def count_episodes(directory): filenames = list(directory.glob('*.npz')) num_episodes = len(filenames) num_steps = sum(int(str(n).split('-')[-1][:-4]) - 1 for n in filenames) return num_episodes, num_steps def save_episode(directory, episode, disable=False): timestamp = datetime.datetime.now().strftime('%Y%m%dT%H%M%S') identifier = str(uuid.uuid4().hex) length = eplen(episode) filename = directory / f'{timestamp}-{identifier}-{length}.npz' if not disable: with io.BytesIO() as f1: np.savez_compressed(f1, **episode) f1.seek(0) with filename.open('wb') as f2: f2.write(f1.read()) return filename def load_episodes(directory, capacity=None, minlen=1, keep_temporal_order=False): # The returned directory from filenames to episodes is guaranteed to be in # temporally sorted order. filenames = sorted(directory.glob('*.npz')) if not keep_temporal_order: print('Shuffling order of offline trajectories!') random.Random(0).shuffle(filenames) if capacity: num_steps = 0 num_episodes = 0 for filename in reversed(filenames): length = int(str(filename).split('-')[-1][:-4]) num_steps += length num_episodes += 1 if num_steps >= capacity: break filenames = filenames[-num_episodes:] episodes = {} for filename in filenames: try: with filename.open('rb') as f: episode = np.load(f) episode = {k: episode[k] for k in episode.keys()} # Conversion for older versions of npz files. if 'is_terminal' not in episode: episode['is_terminal'] = episode['discount'] == 0. except Exception as e: print(f'Could not load episode {str(filename)}: {e}') continue episodes[str(filename)] = episode return episodes def convert(value): value = np.array(value) if np.issubdtype(value.dtype, np.floating): return value.astype(np.float32) elif np.issubdtype(value.dtype, np.signedinteger): return value.astype(np.int32) elif np.issubdtype(value.dtype, np.uint8): return value.astype(np.uint8) return value def eplen(episode): return len(episode['action']) - 1 class OfflineReplay: def __init__( self, directory, capacity=0, ongoing=False, minlen=1, maxlen=0, prioritize_ends=False, split_val=False): self._capacity = capacity self._minlen = minlen self._maxlen = maxlen self._prioritize_ends = prioritize_ends self._random = np.random.RandomState() self._total_episodes = 0 self._total_steps = 0 self._loaded_episodes = 0 self._loaded_steps = 0 self._complete_eps = {} if type(directory) is not list: directory = [directory] for d in directory: path = pathlib.Path(d).expanduser() complete_eps = load_episodes(path, capacity, minlen) self._complete_eps.update(complete_eps) t_eps, t_steps = count_episodes(path) self._total_episodes += t_eps self._total_steps += t_steps self._loaded_episodes += len(complete_eps) self._loaded_steps += sum(eplen(x) for x in complete_eps.values()) print('Loaded from offline directory(ies)!') print(self.stats) self.episodes = list(self._complete_eps.values()) if split_val: self.episodes, self.val_episodes = train_test_split(self.episodes, test_size=0.2) print( f'Split into {len(self.episodes)} training episodes and {len(self.val_episodes)} validation episodes.') @property def stats(self): return { 'total_steps': self._total_steps, 'total_episodes': self._total_episodes, 'loaded_steps': self._loaded_steps, 'loaded_episodes': self._loaded_episodes, } def dataset(self, batch, length, validation=False): example = next(iter(self._generate_chunks(length, validation=validation))) dataset = tf.data.Dataset.from_generator( lambda: self._generate_chunks(length, validation=validation), {k: v.dtype for k, v in example.items()}, {k: v.shape for k, v in example.items()}) dataset = dataset.batch(batch, drop_remainder=True) dataset = dataset.prefetch(5) return dataset def _generate_chunks(self, length, validation): sequence = self._sample_sequence(validation=validation) while True: chunk = collections.defaultdict(list) added = 0 while added < length: needed = length - added adding = {k: v[:needed] for k, v in sequence.items()} sequence = {k: v[needed:] for k, v in sequence.items()} for key, value in adding.items(): chunk[key].append(value) added += len(adding['action']) if len(sequence['action']) < 1: sequence = self._sample_sequence(validation=validation) chunk = {k: np.concatenate(v) for k, v in chunk.items()} yield chunk def _sample_sequence(self, validation): if validation: episode = self._random.choice(self.val_episodes) else: episode = self._random.choice(self.episodes) total = len(episode['action']) length = total if self._maxlen: length = min(length, self._maxlen) # Randomize length to avoid all chunks ending at the same time in case the # episodes are all of the same length. length -= np.random.randint(self._minlen) length = max(self._minlen, length) upper = total - length + 1 if self._prioritize_ends: upper += self._minlen index = min(self._random.randint(upper), total - length) sequence = { k: convert(v[index: index + length]) for k, v in episode.items() if not k.startswith('log_')} sequence['is_first'] = np.zeros(len(sequence['action']), np.bool) sequence['is_first'][0] = True if self._maxlen: assert self._minlen <= len(sequence['action']) <= self._maxlen return sequence # Generate chunks out of the validation episodes. (im loss is infinite...?) def validation_chunks(self, batch, length): chunks = [] for episode in self.val_episodes: episode_chunks = [] sequence = {k: convert(v) for k, v in episode.items() if not k.startswith('log_')} sequence['is_first'] = np.zeros(len(sequence['action']), np.bool) while len(sequence['action']) >= length: chunk = collections.defaultdict(list) adding = {k: v[:length] for k, v in sequence.items()} adding['is_first'][0] = True sequence = {k: v[length:] for k, v in sequence.items()} for key, value in adding.items(): chunk[key].append(value) chunk = {k: np.concatenate(v) for k, v in chunk.items()} episode_chunks.append(chunk) episode_chunks = {k: tf.stack([x[k] for x in episode_chunks]) for k in episode_chunks[0].keys()} chunks.append(episode_chunks) return chunks