File size: 3,083 Bytes
23bc32f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
import math
from typing import Generator, List

import torch

from .batch import Batch
from .episode import Episode
from .segment import Segment, SegmentId


def collate_segments_to_batch(segments: List[Segment]) -> Batch:
    return Batch(
        torch.stack(list(map(lambda s: s.observations, segments))).div(255),
        torch.stack(list(map(lambda s: s.actions, segments))),
        torch.stack(list(map(lambda s: s.rewards, segments))),
        torch.stack(list(map(lambda s: s.ends, segments))),
        torch.stack(list(map(lambda s: s.mask_padding, segments))),
        list(map(lambda segment: segment.id, segments))
    )


def make_segment(episode: Episode, segment_id: SegmentId, should_pad: bool = True) -> Segment:
    assert segment_id.start < len(episode) and segment_id.stop > 0 and segment_id.start < segment_id.stop
    padding_length_right = max(0, segment_id.stop - len(episode))
    padding_length_left = max(0, -segment_id.start)
    assert padding_length_right == padding_length_left == 0 or should_pad

    def pad(x):
        pad_right = torch.nn.functional.pad(x, [0 for _ in range(2 * x.ndim - 1)] + [padding_length_right]) if padding_length_right > 0 else x
        return torch.nn.functional.pad(pad_right, [0 for _ in range(2 * x.ndim - 2)] + [padding_length_left, 0]) if padding_length_left > 0 else pad_right

    start = max(0, segment_id.start)
    stop = min(len(episode), segment_id.stop)

    return Segment(
        pad(episode.observations[start:stop]),
        pad(episode.actions[start:stop]),
        pad(episode.rewards[start:stop]),
        pad(episode.ends[start:stop]),
        mask_padding=torch.cat((torch.zeros(padding_length_left), torch.ones(stop - start), torch.zeros(padding_length_right))).bool(),
        id=SegmentId(segment_id.episode_id, start, stop)
    )


class DatasetTraverser:
    def __init__(self, dataset, batch_num_samples: int, chunk_size: int) -> None:
        self.dataset = dataset
        self.batch_num_samples = batch_num_samples
        self.chunk_size = chunk_size
        self._num_batches = math.ceil(sum([math.ceil(dataset.lengths[episode_id] / chunk_size) - int(dataset.lengths[episode_id] % chunk_size == 1) for episode_id in range(dataset.num_episodes)]) / batch_num_samples)

    def __len__(self) -> int:
        return self._num_batches 

    def __iter__(self) -> Generator[Batch, None, None]:
        chunks = []

        for episode_id in range(self.dataset.num_episodes):
            episode = self.dataset.load_episode(episode_id)
            chunks.extend(make_segment(episode, SegmentId(episode_id, start=i * self.chunk_size, stop=(i + 1) * self.chunk_size), should_pad=True) for i in range(math.ceil(len(episode) / self.chunk_size)))
            if chunks[-1].effective_size < 2:
                chunks.pop()

            while len(chunks) >= self.batch_num_samples:
                yield collate_segments_to_batch(chunks[:self.batch_num_samples])
                chunks = chunks[self.batch_num_samples:]

        if len(chunks) > 0:
            yield collate_segments_to_batch(chunks)