zianrahmad commited on
Commit
1ac1076
·
verified ·
1 Parent(s): f405d22

Upload config.json

Browse files
Files changed (1) hide show
  1. config.json +63 -0
config.json ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "normalization_mapping": {
3
+ "VISUAL": "IDENTITY",
4
+ "STATE": "MIN_MAX",
5
+ "ACTION": "MIN_MAX"
6
+ },
7
+ "input_features": {
8
+ "observation.images.cam_high": {
9
+ "type": "VISUAL",
10
+ "shape": [
11
+ 3,
12
+ 240,
13
+ 320
14
+ ]
15
+ },
16
+ "observation.images.cam_left_wrist": {
17
+ "type": "VISUAL",
18
+ "shape": [
19
+ 3,
20
+ 240,
21
+ 320
22
+ ]
23
+ },
24
+ "observation.images.cam_right_wrist": {
25
+ "type": "VISUAL",
26
+ "shape": [
27
+ 3,
28
+ 240,
29
+ 320
30
+ ]
31
+ },
32
+ "observation.state": {
33
+ "type": "STATE",
34
+ "shape": [
35
+ 14
36
+ ]
37
+ }
38
+ },
39
+ "output_features": {
40
+ "action": {
41
+ "type": "ACTION",
42
+ "shape": [
43
+ 14
44
+ ]
45
+ }
46
+ },
47
+ "device": "cuda",
48
+ "backbone": "Qwen/Qwen3-VL-4B-Instruct",
49
+ "chunk_size": 60,
50
+ "n_action_steps": 60,
51
+ "max_state_dim": 32,
52
+ "max_action_dim": 32,
53
+ "proj_width": 2560,
54
+ "num_steps": 10,
55
+ "attention_implementation": "eager",
56
+ "dit_hidden_size": 1536,
57
+ "dit_num_heads": 32,
58
+ "dit_num_layers": 16,
59
+ "dit_interleave_self_attention": true,
60
+ "dit_cross_attention_dim": 2560,
61
+ "num_noise_per_sample": 1,
62
+ "type": "spirit_vla"
63
+ }