SaiResearch commited on
Commit
10ed9b4
·
verified ·
1 Parent(s): b5f96d7

Upload 15 files

Browse files
iris/atari/mspacman/config.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "ms_pacman_iris_world_model",
3
+ "env": "MsPacmanNoFrameskip-v4",
4
+ "model_type": "iris",
5
+ "metadata": {
6
+ "latent_dim": [1, 16],
7
+ "num_tokens": 340
8
+ },
9
+ "util_folders":{
10
+ "models": "../../src/models"
11
+ },
12
+ "requirements":{
13
+ "-r": "requirements.txt"
14
+ },
15
+ "models": [
16
+ {
17
+ "name": "world_model",
18
+ "framework": null,
19
+ "format": "state_dict",
20
+ "source": {
21
+ "weights_path": "world_model.pt",
22
+ "class_path": "../../src/world_model.py",
23
+ "class_name": "WorldModel",
24
+ "class_args": [
25
+ {
26
+ "vocab_size": 512,
27
+ "act_vocab_size": 9,
28
+ "tokens_per_block": 17,
29
+ "max_blocks": 20,
30
+ "attention": "causal",
31
+ "num_layers": 10,
32
+ "num_heads": 4,
33
+ "embed_dim": 256,
34
+ "embed_pdrop": 0.1,
35
+ "resid_pdrop": 0.1,
36
+ "attn_pdrop": 0.1
37
+ }]
38
+ },
39
+ "signature": {
40
+ "inputs": ["tokens", "past_keys_values"],
41
+ "call_mode": "positional"
42
+ },
43
+ "sub_models":
44
+ [
45
+ {
46
+ "name": "transformer",
47
+ "sub_model_name": "transformer",
48
+ "signature":
49
+ {
50
+ "inputs": ["sequences", "past_keys_values"],
51
+ "call_mode": "positional"
52
+ }
53
+ }
54
+ ],
55
+ "methods":
56
+ [
57
+ {
58
+ "name": "generate_empty_keys_values",
59
+ "method_name": "generate_empty_keys_values",
60
+ "signature":
61
+ {
62
+ "inputs": ["n"]
63
+ }
64
+ }
65
+ ]
66
+ },
67
+ {
68
+ "name": "tokenizer",
69
+ "framework": null,
70
+ "format": "state_dict",
71
+ "source": {
72
+ "weights_path": "tokenizer.pt",
73
+ "class_path": "../../src/tokenizer.py",
74
+ "class_name": "Tokenizer",
75
+ "class_args": [{
76
+ "vocab_size": 512,
77
+ "embed_dim": 512,
78
+ "encoder": {
79
+ "resolution": 64,
80
+ "in_channels": 3,
81
+ "z_channels": 512,
82
+ "ch": 64,
83
+ "ch_mult": [1, 1, 1, 1, 1],
84
+ "num_res_blocks": 2,
85
+ "attn_resolutions": [8, 16],
86
+ "out_ch": 3,
87
+ "dropout": 0.0
88
+ },
89
+ "decoder": {
90
+ "resolution": 64,
91
+ "in_channels": 3,
92
+ "z_channels": 512,
93
+ "ch": 64,
94
+ "ch_mult": [1, 1, 1, 1, 1],
95
+ "num_res_blocks": 2,
96
+ "attn_resolutions": [8, 16],
97
+ "out_ch": 3,
98
+ "dropout": 0.0
99
+ }
100
+ }]
101
+ },
102
+ "signature": {
103
+ "inputs": ["x", "should_preprocess", "should_postprocess"],
104
+ "call_mode": "positional"
105
+ },
106
+ "sub_models":
107
+ [
108
+ {
109
+ "name": "embedding",
110
+ "sub_model_name": "embedding",
111
+ "signature":
112
+ {
113
+ "call_mode": "auto"
114
+ }
115
+ }
116
+ ],
117
+ "methods":
118
+ [
119
+ {
120
+ "name": "decode",
121
+ "method_name": "decode",
122
+ "signature":
123
+ {
124
+ "inputs": ["z", "should_postprocess"]
125
+ }
126
+ },
127
+ {
128
+ "name": "decode_obs_tokens",
129
+ "method_name": "decode_obs_tokens",
130
+ "signature":
131
+ {
132
+ "inputs": ["obs_tokens", "num_observations_tokens"]
133
+ }
134
+ },
135
+ {
136
+ "name": "encode",
137
+ "method_name": "encode",
138
+ "signature":
139
+ {
140
+ "inputs": ["observations", "should_preprocess"]
141
+ }
142
+ }
143
+ ]
144
+ }
145
+ ]
146
+ }
iris/atari/mspacman/tokenizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e497a9cfe6226edd4a8d04dda1719e3ccd87285ace5c042f15f83ccd69cee0ee
3
+ size 75104314
iris/atari/mspacman/world_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c01f323552e98cf9db6a50477a9e8739cbc82d1c5a8eeaefd56593965ea4f6c3
3
+ size 38481598
iris/atari/pong/config.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pong_iris_world_model",
3
+ "env": "PongNoFrameskip-v4",
4
+ "model_type": "iris",
5
+ "metadata": {
6
+ "latent_dim": [1, 16],
7
+ "num_tokens": 340
8
+ },
9
+ "util_folders":{
10
+ "models": "../../src/models"
11
+ },
12
+ "requirements":{
13
+ "-r": "requirements.txt"
14
+ },
15
+ "models": [
16
+ {
17
+ "name": "world_model",
18
+ "framework": null,
19
+ "format": "state_dict",
20
+ "source": {
21
+ "weights_path": "world_model.pt",
22
+ "class_path": "../../src/world_model.py",
23
+ "class_name": "WorldModel",
24
+ "class_args": [
25
+ {
26
+ "vocab_size": 512,
27
+ "act_vocab_size": 6,
28
+ "tokens_per_block": 17,
29
+ "max_blocks": 20,
30
+ "attention": "causal",
31
+ "num_layers": 10,
32
+ "num_heads": 4,
33
+ "embed_dim": 256,
34
+ "embed_pdrop": 0.1,
35
+ "resid_pdrop": 0.1,
36
+ "attn_pdrop": 0.1
37
+ }]
38
+ },
39
+ "signature": {
40
+ "inputs": ["tokens", "past_keys_values"],
41
+ "call_mode": "positional"
42
+ },
43
+ "sub_models":
44
+ [
45
+ {
46
+ "name": "transformer",
47
+ "sub_model_name": "transformer",
48
+ "signature":
49
+ {
50
+ "inputs": ["sequences", "past_keys_values"],
51
+ "call_mode": "positional"
52
+ }
53
+ }
54
+ ],
55
+ "methods":
56
+ [
57
+ {
58
+ "name": "generate_empty_keys_values",
59
+ "method_name": "generate_empty_keys_values",
60
+ "signature":
61
+ {
62
+ "inputs": ["n"]
63
+ }
64
+ }
65
+ ]
66
+ },
67
+ {
68
+ "name": "tokenizer",
69
+ "framework": null,
70
+ "format": "state_dict",
71
+ "source": {
72
+ "weights_path": "tokenizer.pt",
73
+ "class_path": "../../src/tokenizer.py",
74
+ "class_name": "Tokenizer",
75
+ "class_args": [{
76
+ "vocab_size": 512,
77
+ "embed_dim": 512,
78
+ "encoder": {
79
+ "resolution": 64,
80
+ "in_channels": 3,
81
+ "z_channels": 512,
82
+ "ch": 64,
83
+ "ch_mult": [1, 1, 1, 1, 1],
84
+ "num_res_blocks": 2,
85
+ "attn_resolutions": [8, 16],
86
+ "out_ch": 3,
87
+ "dropout": 0.0
88
+ },
89
+ "decoder": {
90
+ "resolution": 64,
91
+ "in_channels": 3,
92
+ "z_channels": 512,
93
+ "ch": 64,
94
+ "ch_mult": [1, 1, 1, 1, 1],
95
+ "num_res_blocks": 2,
96
+ "attn_resolutions": [8, 16],
97
+ "out_ch": 3,
98
+ "dropout": 0.0
99
+ }
100
+ }]
101
+ },
102
+ "signature": {
103
+ "inputs": ["x", "should_preprocess", "should_postprocess"],
104
+ "call_mode": "positional"
105
+ },
106
+ "sub_models":
107
+ [
108
+ {
109
+ "name": "embedding",
110
+ "sub_model_name": "embedding",
111
+ "signature":
112
+ {
113
+ "call_mode": "auto"
114
+ }
115
+ }
116
+ ],
117
+ "methods":
118
+ [
119
+ {
120
+ "name": "decode",
121
+ "method_name": "decode",
122
+ "signature":
123
+ {
124
+ "inputs": ["z", "should_postprocess"]
125
+ }
126
+ },
127
+ {
128
+ "name": "decode_obs_tokens",
129
+ "method_name": "decode_obs_tokens",
130
+ "signature":
131
+ {
132
+ "inputs": ["obs_tokens", "num_observations_tokens"]
133
+ }
134
+ },
135
+ {
136
+ "name": "encode",
137
+ "method_name": "encode",
138
+ "signature":
139
+ {
140
+ "inputs": ["observations", "should_preprocess"]
141
+ }
142
+ }
143
+ ]
144
+ }
145
+ ]
146
+ }
iris/atari/pong/tokenizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f1c533d5c97b9c6edbc450bfcd6b4338e6bd72d961807b41c22d7cfa69c7e63
3
+ size 75104314
iris/atari/pong/world_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:066f3b9617cef56ed5bae3da1b3e563015b6af8b65452ca830d92ae49cf39a83
3
+ size 38478526
iris/atari/privateeye/config.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "private_eye_iris_world_model",
3
+ "env": "PrivateEyeNoFrameskip-v4",
4
+ "model_type": "iris",
5
+ "metadata": {
6
+ "latent_dim": [1, 16],
7
+ "num_tokens": 340
8
+ },
9
+ "util_folders":{
10
+ "models": "../../src/models"
11
+ },
12
+ "requirements":{
13
+ "-r": "requirements.txt"
14
+ },
15
+ "models": [
16
+ {
17
+ "name": "world_model",
18
+ "framework": null,
19
+ "format": "state_dict",
20
+ "source": {
21
+ "weights_path": "world_model.pt",
22
+ "class_path": "../../src/world_model.py",
23
+ "class_name": "WorldModel",
24
+ "class_args": [
25
+ {
26
+ "vocab_size": 512,
27
+ "act_vocab_size": 18,
28
+ "tokens_per_block": 17,
29
+ "max_blocks": 20,
30
+ "attention": "causal",
31
+ "num_layers": 10,
32
+ "num_heads": 4,
33
+ "embed_dim": 256,
34
+ "embed_pdrop": 0.1,
35
+ "resid_pdrop": 0.1,
36
+ "attn_pdrop": 0.1
37
+ }]
38
+ },
39
+ "signature": {
40
+ "inputs": ["tokens", "past_keys_values"],
41
+ "call_mode": "positional"
42
+ },
43
+ "sub_models":
44
+ [
45
+ {
46
+ "name": "transformer",
47
+ "sub_model_name": "transformer",
48
+ "signature":
49
+ {
50
+ "inputs": ["sequences", "past_keys_values"],
51
+ "call_mode": "positional"
52
+ }
53
+ }
54
+ ],
55
+ "methods":
56
+ [
57
+ {
58
+ "name": "generate_empty_keys_values",
59
+ "method_name": "generate_empty_keys_values",
60
+ "signature":
61
+ {
62
+ "inputs": ["n"]
63
+ }
64
+ }
65
+ ]
66
+ },
67
+ {
68
+ "name": "tokenizer",
69
+ "framework": null,
70
+ "format": "state_dict",
71
+ "source": {
72
+ "weights_path": "tokenizer.pt",
73
+ "class_path": "../../src/tokenizer.py",
74
+ "class_name": "Tokenizer",
75
+ "class_args": [{
76
+ "vocab_size": 512,
77
+ "embed_dim": 512,
78
+ "encoder": {
79
+ "resolution": 64,
80
+ "in_channels": 3,
81
+ "z_channels": 512,
82
+ "ch": 64,
83
+ "ch_mult": [1, 1, 1, 1, 1],
84
+ "num_res_blocks": 2,
85
+ "attn_resolutions": [8, 16],
86
+ "out_ch": 3,
87
+ "dropout": 0.0
88
+ },
89
+ "decoder": {
90
+ "resolution": 64,
91
+ "in_channels": 3,
92
+ "z_channels": 512,
93
+ "ch": 64,
94
+ "ch_mult": [1, 1, 1, 1, 1],
95
+ "num_res_blocks": 2,
96
+ "attn_resolutions": [8, 16],
97
+ "out_ch": 3,
98
+ "dropout": 0.0
99
+ }
100
+ }]
101
+ },
102
+ "signature": {
103
+ "inputs": ["x", "should_preprocess", "should_postprocess"],
104
+ "call_mode": "positional"
105
+ },
106
+ "sub_models":
107
+ [
108
+ {
109
+ "name": "embedding",
110
+ "sub_model_name": "embedding",
111
+ "signature":
112
+ {
113
+ "call_mode": "auto"
114
+ }
115
+ }
116
+ ],
117
+ "methods":
118
+ [
119
+ {
120
+ "name": "decode",
121
+ "method_name": "decode",
122
+ "signature":
123
+ {
124
+ "inputs": ["z", "should_postprocess"]
125
+ }
126
+ },
127
+ {
128
+ "name": "decode_obs_tokens",
129
+ "method_name": "decode_obs_tokens",
130
+ "signature":
131
+ {
132
+ "inputs": ["obs_tokens", "num_observations_tokens"]
133
+ }
134
+ },
135
+ {
136
+ "name": "encode",
137
+ "method_name": "encode",
138
+ "signature":
139
+ {
140
+ "inputs": ["observations", "should_preprocess"]
141
+ }
142
+ }
143
+ ]
144
+ }
145
+ ]
146
+ }
iris/atari/privateeye/tokenizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3410a6734c5159a4ce165079e9e0ef0f2311c52a5924c3caf1f266cb801641b8
3
+ size 75104314
iris/atari/privateeye/world_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f88a6f4190283e0674e23815b2c800a7b0faba2b8b8a4948d2d201f13b50542e
3
+ size 38490814
iris/atari/qbert/config.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "qbert_iris_world_model",
3
+ "env": "QbertNoFrameskip-v4",
4
+ "model_type": "iris",
5
+ "metadata": {
6
+ "latent_dim": [1, 16],
7
+ "num_tokens": 340
8
+ },
9
+ "util_folders":{
10
+ "models": "../../src/models"
11
+ },
12
+ "requirements":{
13
+ "-r": "requirements.txt"
14
+ },
15
+ "models": [
16
+ {
17
+ "name": "world_model",
18
+ "framework": null,
19
+ "format": "state_dict",
20
+ "source": {
21
+ "weights_path": "world_model.pt",
22
+ "class_path": "../../src/world_model.py",
23
+ "class_name": "WorldModel",
24
+ "class_args": [
25
+ {
26
+ "vocab_size": 512,
27
+ "act_vocab_size": 6,
28
+ "tokens_per_block": 17,
29
+ "max_blocks": 20,
30
+ "attention": "causal",
31
+ "num_layers": 10,
32
+ "num_heads": 4,
33
+ "embed_dim": 256,
34
+ "embed_pdrop": 0.1,
35
+ "resid_pdrop": 0.1,
36
+ "attn_pdrop": 0.1
37
+ }]
38
+ },
39
+ "signature": {
40
+ "inputs": ["tokens", "past_keys_values"],
41
+ "call_mode": "positional"
42
+ },
43
+ "sub_models":
44
+ [
45
+ {
46
+ "name": "transformer",
47
+ "sub_model_name": "transformer",
48
+ "signature":
49
+ {
50
+ "inputs": ["sequences", "past_keys_values"],
51
+ "call_mode": "positional"
52
+ }
53
+ }
54
+ ],
55
+ "methods":
56
+ [
57
+ {
58
+ "name": "generate_empty_keys_values",
59
+ "method_name": "generate_empty_keys_values",
60
+ "signature":
61
+ {
62
+ "inputs": ["n"]
63
+ }
64
+ }
65
+ ]
66
+ },
67
+ {
68
+ "name": "tokenizer",
69
+ "framework": null,
70
+ "format": "state_dict",
71
+ "source": {
72
+ "weights_path": "tokenizer.pt",
73
+ "class_path": "../../src/tokenizer.py",
74
+ "class_name": "Tokenizer",
75
+ "class_args": [{
76
+ "vocab_size": 512,
77
+ "embed_dim": 512,
78
+ "encoder": {
79
+ "resolution": 64,
80
+ "in_channels": 3,
81
+ "z_channels": 512,
82
+ "ch": 64,
83
+ "ch_mult": [1, 1, 1, 1, 1],
84
+ "num_res_blocks": 2,
85
+ "attn_resolutions": [8, 16],
86
+ "out_ch": 3,
87
+ "dropout": 0.0
88
+ },
89
+ "decoder": {
90
+ "resolution": 64,
91
+ "in_channels": 3,
92
+ "z_channels": 512,
93
+ "ch": 64,
94
+ "ch_mult": [1, 1, 1, 1, 1],
95
+ "num_res_blocks": 2,
96
+ "attn_resolutions": [8, 16],
97
+ "out_ch": 3,
98
+ "dropout": 0.0
99
+ }
100
+ }]
101
+ },
102
+ "signature": {
103
+ "inputs": ["x", "should_preprocess", "should_postprocess"],
104
+ "call_mode": "positional"
105
+ },
106
+ "sub_models":
107
+ [
108
+ {
109
+ "name": "embedding",
110
+ "sub_model_name": "embedding",
111
+ "signature":
112
+ {
113
+ "call_mode": "auto"
114
+ }
115
+ }
116
+ ],
117
+ "methods":
118
+ [
119
+ {
120
+ "name": "decode",
121
+ "method_name": "decode",
122
+ "signature":
123
+ {
124
+ "inputs": ["z", "should_postprocess"]
125
+ }
126
+ },
127
+ {
128
+ "name": "decode_obs_tokens",
129
+ "method_name": "decode_obs_tokens",
130
+ "signature":
131
+ {
132
+ "inputs": ["obs_tokens", "num_observations_tokens"]
133
+ }
134
+ },
135
+ {
136
+ "name": "encode",
137
+ "method_name": "encode",
138
+ "signature":
139
+ {
140
+ "inputs": ["observations", "should_preprocess"]
141
+ }
142
+ }
143
+ ]
144
+ }
145
+ ]
146
+ }
iris/atari/qbert/tokenizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e02cb19d8981d6ff2cc42fe0fc49d0c2f98bcb3e1ef48728395e6135592513b7
3
+ size 75104314
iris/atari/qbert/world_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32a3a0919625cabbcd4f85fdafe046690e798702716993f06cdbeb98ec6b71eb
3
+ size 38478526
iris/atari/roadrunner/config.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "road_runner_iris_world_model",
3
+ "env": "RoadRunnerNoFrameskip-v4",
4
+ "model_type": "iris",
5
+ "metadata": {
6
+ "latent_dim": [1, 16],
7
+ "num_tokens": 340
8
+ },
9
+ "util_folders":{
10
+ "models": "../../src/models"
11
+ },
12
+ "requirements":{
13
+ "-r": "requirements.txt"
14
+ },
15
+ "models": [
16
+ {
17
+ "name": "world_model",
18
+ "framework": null,
19
+ "format": "state_dict",
20
+ "source": {
21
+ "weights_path": "world_model.pt",
22
+ "class_path": "../../src/world_model.py",
23
+ "class_name": "WorldModel",
24
+ "class_args": [
25
+ {
26
+ "vocab_size": 512,
27
+ "act_vocab_size": 18,
28
+ "tokens_per_block": 17,
29
+ "max_blocks": 20,
30
+ "attention": "causal",
31
+ "num_layers": 10,
32
+ "num_heads": 4,
33
+ "embed_dim": 256,
34
+ "embed_pdrop": 0.1,
35
+ "resid_pdrop": 0.1,
36
+ "attn_pdrop": 0.1
37
+ }]
38
+ },
39
+ "signature": {
40
+ "inputs": ["tokens", "past_keys_values"],
41
+ "call_mode": "positional"
42
+ },
43
+ "sub_models":
44
+ [
45
+ {
46
+ "name": "transformer",
47
+ "sub_model_name": "transformer",
48
+ "signature":
49
+ {
50
+ "inputs": ["sequences", "past_keys_values"],
51
+ "call_mode": "positional"
52
+ }
53
+ }
54
+ ],
55
+ "methods":
56
+ [
57
+ {
58
+ "name": "generate_empty_keys_values",
59
+ "method_name": "generate_empty_keys_values",
60
+ "signature":
61
+ {
62
+ "inputs": ["n"]
63
+ }
64
+ }
65
+ ]
66
+ },
67
+ {
68
+ "name": "tokenizer",
69
+ "framework": null,
70
+ "format": "state_dict",
71
+ "source": {
72
+ "weights_path": "tokenizer.pt",
73
+ "class_path": "../../src/tokenizer.py",
74
+ "class_name": "Tokenizer",
75
+ "class_args": [{
76
+ "vocab_size": 512,
77
+ "embed_dim": 512,
78
+ "encoder": {
79
+ "resolution": 64,
80
+ "in_channels": 3,
81
+ "z_channels": 512,
82
+ "ch": 64,
83
+ "ch_mult": [1, 1, 1, 1, 1],
84
+ "num_res_blocks": 2,
85
+ "attn_resolutions": [8, 16],
86
+ "out_ch": 3,
87
+ "dropout": 0.0
88
+ },
89
+ "decoder": {
90
+ "resolution": 64,
91
+ "in_channels": 3,
92
+ "z_channels": 512,
93
+ "ch": 64,
94
+ "ch_mult": [1, 1, 1, 1, 1],
95
+ "num_res_blocks": 2,
96
+ "attn_resolutions": [8, 16],
97
+ "out_ch": 3,
98
+ "dropout": 0.0
99
+ }
100
+ }]
101
+ },
102
+ "signature": {
103
+ "inputs": ["x", "should_preprocess", "should_postprocess"],
104
+ "call_mode": "positional"
105
+ },
106
+ "sub_models":
107
+ [
108
+ {
109
+ "name": "embedding",
110
+ "sub_model_name": "embedding",
111
+ "signature":
112
+ {
113
+ "call_mode": "auto"
114
+ }
115
+ }
116
+ ],
117
+ "methods":
118
+ [
119
+ {
120
+ "name": "decode",
121
+ "method_name": "decode",
122
+ "signature":
123
+ {
124
+ "inputs": ["z", "should_postprocess"]
125
+ }
126
+ },
127
+ {
128
+ "name": "decode_obs_tokens",
129
+ "method_name": "decode_obs_tokens",
130
+ "signature":
131
+ {
132
+ "inputs": ["obs_tokens", "num_observations_tokens"]
133
+ }
134
+ },
135
+ {
136
+ "name": "encode",
137
+ "method_name": "encode",
138
+ "signature":
139
+ {
140
+ "inputs": ["observations", "should_preprocess"]
141
+ }
142
+ }
143
+ ]
144
+ }
145
+ ]
146
+ }
iris/atari/roadrunner/tokenizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0792e4df0fbec0e02da4f11259228d8c48f6de6ac0756eda702ca988ab58da09
3
+ size 75104314
iris/atari/roadrunner/world_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91fdd70012d26d4f8bd632554d663fc49cc263681bee6fb4c33efe601b2d6ce2
3
+ size 38490814