Robotics
Safetensors
vision-language-action-model
Jia-Zeng commited on
Commit
08cf716
·
verified ·
1 Parent(s): 19e953f

Update checkpoint: pretrained on heterogeneous data sources over real-robot data, synthetic simulation data, and human videos.

Browse files
Files changed (5) hide show
  1. README.md +1 -1
  2. config.json +3 -3
  3. model.safetensors +1 -1
  4. stats.json +0 -0
  5. train_config.json +0 -0
README.md CHANGED
@@ -25,7 +25,7 @@ datasets:
25
 
26
  Building upon InternVL3 and Qwen3-VL, we instantiate InternVLA-A1 at 2B and 3B parameter scales. Covering different model scales and pre-training data configurations, we release the InternVLA-A1 series:
27
 
28
- - [x] [InternVLA-A1-3B](https://huggingface.co/InternRobotics/InternVLA-A1-3B): pretrained on the large-scale, high-fidelity simulation data [InternData-A1](https://huggingface.co/datasets/InternRobotics/InternData-A1), together with open-source robot data (e.g. Agibot-World)
29
  - [x] [InternVLA-A1-3B-RoboTwin](https://huggingface.co/InternRobotics/InternVLA-A1-3B-RoboTwin): finetuned on RoboTwin 2.0 benchmark
30
  - [ ] [InternVLA-A1-3B-Pretrain-InternData-A1](https://huggingface.co/InternRobotics/InternVLA-A1-3B-Pretrain-InternData-A1): pretrained on InternData-A1 only
31
  - [ ] [InternVLA-A1-2B-Pretrain-InternData-A1](https://huggingface.co/InternRobotics/InternVLA-A1-2B-Pretrain-InternData-A1): pretrained on InternData-A1 only
 
25
 
26
  Building upon InternVL3 and Qwen3-VL, we instantiate InternVLA-A1 at 2B and 3B parameter scales. Covering different model scales and pre-training data configurations, we release the InternVLA-A1 series:
27
 
28
+ - [x] [InternVLA-A1-3B](https://huggingface.co/InternRobotics/InternVLA-A1-3B): pretrained on heterogeneous data sources over real-robot data, synthetic simulation data, and human videos
29
  - [x] [InternVLA-A1-3B-RoboTwin](https://huggingface.co/InternRobotics/InternVLA-A1-3B-RoboTwin): finetuned on RoboTwin 2.0 benchmark
30
  - [ ] [InternVLA-A1-3B-Pretrain-InternData-A1](https://huggingface.co/InternRobotics/InternVLA-A1-3B-Pretrain-InternData-A1): pretrained on InternData-A1 only
31
  - [ ] [InternVLA-A1-2B-Pretrain-InternData-A1](https://huggingface.co/InternRobotics/InternVLA-A1-2B-Pretrain-InternData-A1): pretrained on InternData-A1 only
config.json CHANGED
@@ -20,11 +20,11 @@
20
  "device": "cuda",
21
  "use_amp": false,
22
  "push_to_hub": false,
23
- "repo_id": "jcaiaq/qwena1",
24
  "private": null,
25
  "tags": null,
26
  "license": null,
27
- "pretrained_path": null,
28
  "qwen3_vl_variant": "qwen3_vl_28l",
29
  "action_expert_variant": "qwen3_28l",
30
  "dtype": "bfloat16",
@@ -61,7 +61,7 @@
61
  "optimizer_weight_decay": 0.01,
62
  "optimizer_grad_clip_norm": 1.0,
63
  "scheduler_warmup_steps": 0,
64
- "scheduler_decay_steps": 700000,
65
  "scheduler_decay_lr": 5e-05,
66
  "tokenizer_max_length": 48,
67
  "freeze_vision_encoder": false,
 
20
  "device": "cuda",
21
  "use_amp": false,
22
  "push_to_hub": false,
23
+ "repo_id": "lerobot_lab/qwena1",
24
  "private": null,
25
  "tags": null,
26
  "license": null,
27
+ "pretrained_path": "outputs/qwena1/2026_01_26_13_29_09-qwena1-a1_robotwin_egodex_agibotworld-delta-pretrain-agibotworld_wo_act_loss/checkpoints/01310000/pretrained_model",
28
  "qwen3_vl_variant": "qwen3_vl_28l",
29
  "action_expert_variant": "qwen3_28l",
30
  "dtype": "bfloat16",
 
61
  "optimizer_weight_decay": 0.01,
62
  "optimizer_grad_clip_norm": 1.0,
63
  "scheduler_warmup_steps": 0,
64
+ "scheduler_decay_steps": 25000000,
65
  "scheduler_decay_lr": 5e-05,
66
  "tokenizer_max_length": 48,
67
  "freeze_vision_encoder": false,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c4dfbc04b9783c9e27e492c1df3379292e75884eb9b019b66a73872daa6afdb
3
  size 6721209208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41b0fc1e6772257ab5c47f46338e8438cb8c05ef0a6f2c83123b1a614c397f59
3
  size 6721209208
stats.json CHANGED
The diff for this file is too large to render. See raw diff
 
train_config.json CHANGED
The diff for this file is too large to render. See raw diff