SaiResearch commited on
Commit
798ca0f
·
verified ·
1 Parent(s): 126abf5

Upload 12 files

Browse files
iris/atari/bankheist/config.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "bankheist_iris_world_model",
3
+ "env": "BankHeistNoFrameskip-v4",
4
+ "model_type": "iris",
5
+ "metadata": {
6
+ "latent_dim": [1, 16],
7
+ "num_tokens": 340
8
+ },
9
+ "util_folders":{
10
+ "models": "../../src/models"
11
+ },
12
+ "requirements":{
13
+ "-r": "requirements.txt"
14
+ },
15
+ "models": [
16
+ {
17
+ "name": "world_model",
18
+ "framework": null,
19
+ "format": "state_dict",
20
+ "source": {
21
+ "weights_path": "world_model.pt",
22
+ "class_path": "../../src/world_model.py",
23
+ "class_name": "WorldModel",
24
+ "class_args": [
25
+ {
26
+ "vocab_size": 512,
27
+ "act_vocab_size": 18,
28
+ "tokens_per_block": 17,
29
+ "max_blocks": 20,
30
+ "attention": "causal",
31
+ "num_layers": 10,
32
+ "num_heads": 4,
33
+ "embed_dim": 256,
34
+ "embed_pdrop": 0.1,
35
+ "resid_pdrop": 0.1,
36
+ "attn_pdrop": 0.1
37
+ }]
38
+ },
39
+ "signature": {
40
+ "inputs": ["tokens", "past_keys_values"],
41
+ "call_mode": "positional"
42
+ },
43
+ "sub_models":
44
+ [
45
+ {
46
+ "name": "transformer",
47
+ "sub_model_name": "transformer",
48
+ "signature":
49
+ {
50
+ "inputs": ["sequences", "past_keys_values"],
51
+ "call_mode": "positional"
52
+ }
53
+ }
54
+ ],
55
+ "methods":
56
+ [
57
+ {
58
+ "name": "generate_empty_keys_values",
59
+ "method_name": "generate_empty_keys_values",
60
+ "signature":
61
+ {
62
+ "inputs": ["n"]
63
+ }
64
+ }
65
+ ]
66
+ },
67
+ {
68
+ "name": "tokenizer",
69
+ "framework": null,
70
+ "format": "state_dict",
71
+ "source": {
72
+ "weights_path": "tokenizer.pt",
73
+ "class_path": "../../src/tokenizer.py",
74
+ "class_name": "Tokenizer",
75
+ "class_args": [{
76
+ "vocab_size": 512,
77
+ "embed_dim": 512,
78
+ "encoder": {
79
+ "resolution": 64,
80
+ "in_channels": 3,
81
+ "z_channels": 512,
82
+ "ch": 64,
83
+ "ch_mult": [1, 1, 1, 1, 1],
84
+ "num_res_blocks": 2,
85
+ "attn_resolutions": [8, 16],
86
+ "out_ch": 3,
87
+ "dropout": 0.0
88
+ },
89
+ "decoder": {
90
+ "resolution": 64,
91
+ "in_channels": 3,
92
+ "z_channels": 512,
93
+ "ch": 64,
94
+ "ch_mult": [1, 1, 1, 1, 1],
95
+ "num_res_blocks": 2,
96
+ "attn_resolutions": [8, 16],
97
+ "out_ch": 3,
98
+ "dropout": 0.0
99
+ }
100
+ }]
101
+ },
102
+ "signature": {
103
+ "inputs": ["x", "should_preprocess", "should_postprocess"],
104
+ "call_mode": "positional"
105
+ },
106
+ "sub_models":
107
+ [
108
+ {
109
+ "name": "embedding",
110
+ "sub_model_name": "embedding",
111
+ "signature":
112
+ {
113
+ "call_mode": "auto"
114
+ }
115
+ }
116
+ ],
117
+ "methods":
118
+ [
119
+ {
120
+ "name": "decode",
121
+ "method_name": "decode",
122
+ "signature":
123
+ {
124
+ "inputs": ["z", "should_postprocess"]
125
+ }
126
+ },
127
+ {
128
+ "name": "decode_obs_tokens",
129
+ "method_name": "decode_obs_tokens",
130
+ "signature":
131
+ {
132
+ "inputs": ["obs_tokens", "num_observations_tokens"]
133
+ }
134
+ },
135
+ {
136
+ "name": "encode",
137
+ "method_name": "encode",
138
+ "signature":
139
+ {
140
+ "inputs": ["observations", "should_preprocess"]
141
+ }
142
+ }
143
+ ]
144
+ }
145
+ ]
146
+ }
iris/atari/bankheist/tokenizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae0e59bd9e87f0f39faeb3395a20033f75a9d5632f352f5c63f72d1e0db22204
3
+ size 75104314
iris/atari/bankheist/world_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e81d04acf4f11a003340eb158180a2b6b2f45fee80b8e6876f24502c94b2680f
3
+ size 38490814
iris/atari/battlezone/config.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "battlezone_iris_world_model",
3
+ "env": "BattleZoneNoFrameskip-v4",
4
+ "model_type": "iris",
5
+ "metadata": {
6
+ "latent_dim": [1, 16],
7
+ "num_tokens": 340
8
+ },
9
+ "util_folders":{
10
+ "models": "../../src/models"
11
+ },
12
+ "requirements":{
13
+ "-r": "requirements.txt"
14
+ },
15
+ "models": [
16
+ {
17
+ "name": "world_model",
18
+ "framework": null,
19
+ "format": "state_dict",
20
+ "source": {
21
+ "weights_path": "world_model.pt",
22
+ "class_path": "../../src/world_model.py",
23
+ "class_name": "WorldModel",
24
+ "class_args": [
25
+ {
26
+ "vocab_size": 512,
27
+ "act_vocab_size": 18,
28
+ "tokens_per_block": 17,
29
+ "max_blocks": 20,
30
+ "attention": "causal",
31
+ "num_layers": 10,
32
+ "num_heads": 4,
33
+ "embed_dim": 256,
34
+ "embed_pdrop": 0.1,
35
+ "resid_pdrop": 0.1,
36
+ "attn_pdrop": 0.1
37
+ }]
38
+ },
39
+ "signature": {
40
+ "inputs": ["tokens", "past_keys_values"],
41
+ "call_mode": "positional"
42
+ },
43
+ "sub_models":
44
+ [
45
+ {
46
+ "name": "transformer",
47
+ "sub_model_name": "transformer",
48
+ "signature":
49
+ {
50
+ "inputs": ["sequences", "past_keys_values"],
51
+ "call_mode": "positional"
52
+ }
53
+ }
54
+ ],
55
+ "methods":
56
+ [
57
+ {
58
+ "name": "generate_empty_keys_values",
59
+ "method_name": "generate_empty_keys_values",
60
+ "signature":
61
+ {
62
+ "inputs": ["n"]
63
+ }
64
+ }
65
+ ]
66
+ },
67
+ {
68
+ "name": "tokenizer",
69
+ "framework": null,
70
+ "format": "state_dict",
71
+ "source": {
72
+ "weights_path": "tokenizer.pt",
73
+ "class_path": "../../src/tokenizer.py",
74
+ "class_name": "Tokenizer",
75
+ "class_args": [{
76
+ "vocab_size": 512,
77
+ "embed_dim": 512,
78
+ "encoder": {
79
+ "resolution": 64,
80
+ "in_channels": 3,
81
+ "z_channels": 512,
82
+ "ch": 64,
83
+ "ch_mult": [1, 1, 1, 1, 1],
84
+ "num_res_blocks": 2,
85
+ "attn_resolutions": [8, 16],
86
+ "out_ch": 3,
87
+ "dropout": 0.0
88
+ },
89
+ "decoder": {
90
+ "resolution": 64,
91
+ "in_channels": 3,
92
+ "z_channels": 512,
93
+ "ch": 64,
94
+ "ch_mult": [1, 1, 1, 1, 1],
95
+ "num_res_blocks": 2,
96
+ "attn_resolutions": [8, 16],
97
+ "out_ch": 3,
98
+ "dropout": 0.0
99
+ }
100
+ }]
101
+ },
102
+ "signature": {
103
+ "inputs": ["x", "should_preprocess", "should_postprocess"],
104
+ "call_mode": "positional"
105
+ },
106
+ "sub_models":
107
+ [
108
+ {
109
+ "name": "embedding",
110
+ "sub_model_name": "embedding",
111
+ "signature":
112
+ {
113
+ "call_mode": "auto"
114
+ }
115
+ }
116
+ ],
117
+ "methods":
118
+ [
119
+ {
120
+ "name": "decode",
121
+ "method_name": "decode",
122
+ "signature":
123
+ {
124
+ "inputs": ["z", "should_postprocess"]
125
+ }
126
+ },
127
+ {
128
+ "name": "decode_obs_tokens",
129
+ "method_name": "decode_obs_tokens",
130
+ "signature":
131
+ {
132
+ "inputs": ["obs_tokens", "num_observations_tokens"]
133
+ }
134
+ },
135
+ {
136
+ "name": "encode",
137
+ "method_name": "encode",
138
+ "signature":
139
+ {
140
+ "inputs": ["observations", "should_preprocess"]
141
+ }
142
+ }
143
+ ]
144
+ }
145
+ ]
146
+ }
iris/atari/battlezone/tokenizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:903ecacfc9a8b2328f38ee3a6afc74c0724fb16949ecddc53648ca925be8e12c
3
+ size 75104314
iris/atari/battlezone/world_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2abb50dfd0e5adef304544da9f3951c7f4a3b976e9eb76096fea8969842b3db1
3
+ size 38490814
iris/atari/boxing/config.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "boxing_iris_world_model",
3
+ "env": "BoxingNoFrameskip-v4",
4
+ "model_type": "iris",
5
+ "metadata": {
6
+ "latent_dim": [1, 16],
7
+ "num_tokens": 340
8
+ },
9
+ "util_folders":{
10
+ "models": "../../src/models"
11
+ },
12
+ "requirements":{
13
+ "-r": "requirements.txt"
14
+ },
15
+ "models": [
16
+ {
17
+ "name": "world_model",
18
+ "framework": null,
19
+ "format": "state_dict",
20
+ "source": {
21
+ "weights_path": "world_model.pt",
22
+ "class_path": "../../src/world_model.py",
23
+ "class_name": "WorldModel",
24
+ "class_args": [
25
+ {
26
+ "vocab_size": 512,
27
+ "act_vocab_size": 18,
28
+ "tokens_per_block": 17,
29
+ "max_blocks": 20,
30
+ "attention": "causal",
31
+ "num_layers": 10,
32
+ "num_heads": 4,
33
+ "embed_dim": 256,
34
+ "embed_pdrop": 0.1,
35
+ "resid_pdrop": 0.1,
36
+ "attn_pdrop": 0.1
37
+ }]
38
+ },
39
+ "signature": {
40
+ "inputs": ["tokens", "past_keys_values"],
41
+ "call_mode": "positional"
42
+ },
43
+ "sub_models":
44
+ [
45
+ {
46
+ "name": "transformer",
47
+ "sub_model_name": "transformer",
48
+ "signature":
49
+ {
50
+ "inputs": ["sequences", "past_keys_values"],
51
+ "call_mode": "positional"
52
+ }
53
+ }
54
+ ],
55
+ "methods":
56
+ [
57
+ {
58
+ "name": "generate_empty_keys_values",
59
+ "method_name": "generate_empty_keys_values",
60
+ "signature":
61
+ {
62
+ "inputs": ["n"]
63
+ }
64
+ }
65
+ ]
66
+ },
67
+ {
68
+ "name": "tokenizer",
69
+ "framework": null,
70
+ "format": "state_dict",
71
+ "source": {
72
+ "weights_path": "tokenizer.pt",
73
+ "class_path": "../../src/tokenizer.py",
74
+ "class_name": "Tokenizer",
75
+ "class_args": [{
76
+ "vocab_size": 512,
77
+ "embed_dim": 512,
78
+ "encoder": {
79
+ "resolution": 64,
80
+ "in_channels": 3,
81
+ "z_channels": 512,
82
+ "ch": 64,
83
+ "ch_mult": [1, 1, 1, 1, 1],
84
+ "num_res_blocks": 2,
85
+ "attn_resolutions": [8, 16],
86
+ "out_ch": 3,
87
+ "dropout": 0.0
88
+ },
89
+ "decoder": {
90
+ "resolution": 64,
91
+ "in_channels": 3,
92
+ "z_channels": 512,
93
+ "ch": 64,
94
+ "ch_mult": [1, 1, 1, 1, 1],
95
+ "num_res_blocks": 2,
96
+ "attn_resolutions": [8, 16],
97
+ "out_ch": 3,
98
+ "dropout": 0.0
99
+ }
100
+ }]
101
+ },
102
+ "signature": {
103
+ "inputs": ["x", "should_preprocess", "should_postprocess"],
104
+ "call_mode": "positional"
105
+ },
106
+ "sub_models":
107
+ [
108
+ {
109
+ "name": "embedding",
110
+ "sub_model_name": "embedding",
111
+ "signature":
112
+ {
113
+ "call_mode": "auto"
114
+ }
115
+ }
116
+ ],
117
+ "methods":
118
+ [
119
+ {
120
+ "name": "decode",
121
+ "method_name": "decode",
122
+ "signature":
123
+ {
124
+ "inputs": ["z", "should_postprocess"]
125
+ }
126
+ },
127
+ {
128
+ "name": "decode_obs_tokens",
129
+ "method_name": "decode_obs_tokens",
130
+ "signature":
131
+ {
132
+ "inputs": ["obs_tokens", "num_observations_tokens"]
133
+ }
134
+ },
135
+ {
136
+ "name": "encode",
137
+ "method_name": "encode",
138
+ "signature":
139
+ {
140
+ "inputs": ["observations", "should_preprocess"]
141
+ }
142
+ }
143
+ ]
144
+ }
145
+ ]
146
+ }
iris/atari/boxing/tokenizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c603d6ba26f15ff677945d6065a109e48c62487bedef1b089013e6ec7c0e7e4
3
+ size 75104314
iris/atari/boxing/world_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d88fe6f63f20745a16f2d9a38e4ea1e2584ead0e10c87f41a01aa8e3785f9485
3
+ size 38490814
iris/atari/breakout/config.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "breakout_iris_world_model",
3
+ "env": "BreakoutNoFrameskip-v4",
4
+ "model_type": "iris",
5
+ "metadata": {
6
+ "latent_dim": [1, 16],
7
+ "num_tokens": 340
8
+ },
9
+ "util_folders":{
10
+ "models": "../../src/models"
11
+ },
12
+ "requirements":{
13
+ "-r": "requirements.txt"
14
+ },
15
+ "models": [
16
+ {
17
+ "name": "world_model",
18
+ "framework": null,
19
+ "format": "state_dict",
20
+ "source": {
21
+ "weights_path": "world_model.pt",
22
+ "class_path": "../../src/world_model.py",
23
+ "class_name": "WorldModel",
24
+ "class_args": [
25
+ {
26
+ "vocab_size": 512,
27
+ "act_vocab_size": 4,
28
+ "tokens_per_block": 17,
29
+ "max_blocks": 20,
30
+ "attention": "causal",
31
+ "num_layers": 10,
32
+ "num_heads": 4,
33
+ "embed_dim": 256,
34
+ "embed_pdrop": 0.1,
35
+ "resid_pdrop": 0.1,
36
+ "attn_pdrop": 0.1
37
+ }]
38
+ },
39
+ "signature": {
40
+ "inputs": ["tokens", "past_keys_values"],
41
+ "call_mode": "positional"
42
+ },
43
+ "sub_models":
44
+ [
45
+ {
46
+ "name": "transformer",
47
+ "sub_model_name": "transformer",
48
+ "signature":
49
+ {
50
+ "inputs": ["sequences", "past_keys_values"],
51
+ "call_mode": "positional"
52
+ }
53
+ }
54
+ ],
55
+ "methods":
56
+ [
57
+ {
58
+ "name": "generate_empty_keys_values",
59
+ "method_name": "generate_empty_keys_values",
60
+ "signature":
61
+ {
62
+ "inputs": ["n"]
63
+ }
64
+ }
65
+ ]
66
+ },
67
+ {
68
+ "name": "tokenizer",
69
+ "framework": null,
70
+ "format": "state_dict",
71
+ "source": {
72
+ "weights_path": "tokenizer.pt",
73
+ "class_path": "../../src/tokenizer.py",
74
+ "class_name": "Tokenizer",
75
+ "class_args": [{
76
+ "vocab_size": 512,
77
+ "embed_dim": 512,
78
+ "encoder": {
79
+ "resolution": 64,
80
+ "in_channels": 3,
81
+ "z_channels": 512,
82
+ "ch": 64,
83
+ "ch_mult": [1, 1, 1, 1, 1],
84
+ "num_res_blocks": 2,
85
+ "attn_resolutions": [8, 16],
86
+ "out_ch": 3,
87
+ "dropout": 0.0
88
+ },
89
+ "decoder": {
90
+ "resolution": 64,
91
+ "in_channels": 3,
92
+ "z_channels": 512,
93
+ "ch": 64,
94
+ "ch_mult": [1, 1, 1, 1, 1],
95
+ "num_res_blocks": 2,
96
+ "attn_resolutions": [8, 16],
97
+ "out_ch": 3,
98
+ "dropout": 0.0
99
+ }
100
+ }]
101
+ },
102
+ "signature": {
103
+ "inputs": ["x", "should_preprocess", "should_postprocess"],
104
+ "call_mode": "positional"
105
+ },
106
+ "sub_models":
107
+ [
108
+ {
109
+ "name": "embedding",
110
+ "sub_model_name": "embedding",
111
+ "signature":
112
+ {
113
+ "call_mode": "auto"
114
+ }
115
+ }
116
+ ],
117
+ "methods":
118
+ [
119
+ {
120
+ "name": "decode",
121
+ "method_name": "decode",
122
+ "signature":
123
+ {
124
+ "inputs": ["z", "should_postprocess"]
125
+ }
126
+ },
127
+ {
128
+ "name": "decode_obs_tokens",
129
+ "method_name": "decode_obs_tokens",
130
+ "signature":
131
+ {
132
+ "inputs": ["obs_tokens", "num_observations_tokens"]
133
+ }
134
+ },
135
+ {
136
+ "name": "encode",
137
+ "method_name": "encode",
138
+ "signature":
139
+ {
140
+ "inputs": ["observations", "should_preprocess"]
141
+ }
142
+ }
143
+ ]
144
+ }
145
+ ]
146
+ }
iris/atari/breakout/tokenizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a702fa66a36fabac44c2f83aca92c8dd4130b7774fa82434b2ed48062720005e
3
+ size 75104314
iris/atari/breakout/world_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1714bcc2022fb6367629f7e39f746e4708888103b30b8c0a28b5202e729d419
3
+ size 38476478