Update checkpoint: pretrained on heterogeneous data sources over real-robot data, synthetic simulation data, and human videos.
Browse files- README.md +1 -1
- config.json +3 -3
- model.safetensors +1 -1
- stats.json +0 -0
- train_config.json +0 -0
README.md
CHANGED
|
@@ -25,7 +25,7 @@ datasets:
|
|
| 25 |
|
| 26 |
Building upon InternVL3 and Qwen3-VL, we instantiate InternVLA-A1 at 2B and 3B parameter scales. Covering different model scales and pre-training data configurations, we release the InternVLA-A1 series:
|
| 27 |
|
| 28 |
-
- [x] [InternVLA-A1-3B](https://huggingface.co/InternRobotics/InternVLA-A1-3B): pretrained on
|
| 29 |
- [x] [InternVLA-A1-3B-RoboTwin](https://huggingface.co/InternRobotics/InternVLA-A1-3B-RoboTwin): finetuned on RoboTwin 2.0 benchmark
|
| 30 |
- [ ] [InternVLA-A1-3B-Pretrain-InternData-A1](https://huggingface.co/InternRobotics/InternVLA-A1-3B-Pretrain-InternData-A1): pretrained on InternData-A1 only
|
| 31 |
- [ ] [InternVLA-A1-2B-Pretrain-InternData-A1](https://huggingface.co/InternRobotics/InternVLA-A1-2B-Pretrain-InternData-A1): pretrained on InternData-A1 only
|
|
|
|
| 25 |
|
| 26 |
Building upon InternVL3 and Qwen3-VL, we instantiate InternVLA-A1 at 2B and 3B parameter scales. Covering different model scales and pre-training data configurations, we release the InternVLA-A1 series:
|
| 27 |
|
| 28 |
+
- [x] [InternVLA-A1-3B](https://huggingface.co/InternRobotics/InternVLA-A1-3B): pretrained on heterogeneous data sources over real-robot data, synthetic simulation data, and human videos
|
| 29 |
- [x] [InternVLA-A1-3B-RoboTwin](https://huggingface.co/InternRobotics/InternVLA-A1-3B-RoboTwin): finetuned on RoboTwin 2.0 benchmark
|
| 30 |
- [ ] [InternVLA-A1-3B-Pretrain-InternData-A1](https://huggingface.co/InternRobotics/InternVLA-A1-3B-Pretrain-InternData-A1): pretrained on InternData-A1 only
|
| 31 |
- [ ] [InternVLA-A1-2B-Pretrain-InternData-A1](https://huggingface.co/InternRobotics/InternVLA-A1-2B-Pretrain-InternData-A1): pretrained on InternData-A1 only
|
config.json
CHANGED
|
@@ -20,11 +20,11 @@
|
|
| 20 |
"device": "cuda",
|
| 21 |
"use_amp": false,
|
| 22 |
"push_to_hub": false,
|
| 23 |
-
"repo_id": "
|
| 24 |
"private": null,
|
| 25 |
"tags": null,
|
| 26 |
"license": null,
|
| 27 |
-
"pretrained_path":
|
| 28 |
"qwen3_vl_variant": "qwen3_vl_28l",
|
| 29 |
"action_expert_variant": "qwen3_28l",
|
| 30 |
"dtype": "bfloat16",
|
|
@@ -61,7 +61,7 @@
|
|
| 61 |
"optimizer_weight_decay": 0.01,
|
| 62 |
"optimizer_grad_clip_norm": 1.0,
|
| 63 |
"scheduler_warmup_steps": 0,
|
| 64 |
-
"scheduler_decay_steps":
|
| 65 |
"scheduler_decay_lr": 5e-05,
|
| 66 |
"tokenizer_max_length": 48,
|
| 67 |
"freeze_vision_encoder": false,
|
|
|
|
| 20 |
"device": "cuda",
|
| 21 |
"use_amp": false,
|
| 22 |
"push_to_hub": false,
|
| 23 |
+
"repo_id": "lerobot_lab/qwena1",
|
| 24 |
"private": null,
|
| 25 |
"tags": null,
|
| 26 |
"license": null,
|
| 27 |
+
"pretrained_path": "outputs/qwena1/2026_01_26_13_29_09-qwena1-a1_robotwin_egodex_agibotworld-delta-pretrain-agibotworld_wo_act_loss/checkpoints/01310000/pretrained_model",
|
| 28 |
"qwen3_vl_variant": "qwen3_vl_28l",
|
| 29 |
"action_expert_variant": "qwen3_28l",
|
| 30 |
"dtype": "bfloat16",
|
|
|
|
| 61 |
"optimizer_weight_decay": 0.01,
|
| 62 |
"optimizer_grad_clip_norm": 1.0,
|
| 63 |
"scheduler_warmup_steps": 0,
|
| 64 |
+
"scheduler_decay_steps": 25000000,
|
| 65 |
"scheduler_decay_lr": 5e-05,
|
| 66 |
"tokenizer_max_length": 48,
|
| 67 |
"freeze_vision_encoder": false,
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 6721209208
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:41b0fc1e6772257ab5c47f46338e8438cb8c05ef0a6f2c83123b1a614c397f59
|
| 3 |
size 6721209208
|
stats.json
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train_config.json
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|