| { | |
| "chunk_size": 100, | |
| "freeze_pretrained": true, | |
| "input_shapes": { | |
| "observation.images.phone": [ | |
| 3, | |
| 480, | |
| 640 | |
| ], | |
| "observation.state": [ | |
| 6 | |
| ] | |
| }, | |
| "n_action_steps": 100, | |
| "n_obs_steps": 1, | |
| "num_image_tokens": 729, | |
| "open_lm_config": [ | |
| "--model", | |
| "79m", | |
| "--torchcompile", | |
| "--qk-norm" | |
| ], | |
| "output_shapes": { | |
| "action": [ | |
| 6 | |
| ] | |
| }, | |
| "pretrained_backbone_weights": null, | |
| "tokenizer_name": "EleutherAI/gpt-neox-20b", | |
| "vision_backbone": "siglip" | |
| } |