Update README.md
Browse files
README.md
CHANGED
|
@@ -1,5 +1,4 @@
|
|
| 1 |
---
|
| 2 |
-
license: mit
|
| 3 |
library_name: pytorch
|
| 4 |
tags:
|
| 5 |
- robotics
|
|
@@ -31,31 +30,13 @@ This model is trained for robotic manipulation tasks using vision-language-actio
|
|
| 31 |
- **Dataset**: LIBERO-10 (29 subtasks, 1,354 demonstrations)
|
| 32 |
- **Segmentation**: Semantic action chunking using Gemini Vision API
|
| 33 |
- **Framework**: PyTorch
|
| 34 |
-
- **Checkpoint**: Epoch 90
|
| 35 |
-
|
| 36 |
-
## Usage
|
| 37 |
-
|
| 38 |
-
```python
|
| 39 |
-
import torch
|
| 40 |
-
from pathlib import Path
|
| 41 |
-
|
| 42 |
-
# Load checkpoint
|
| 43 |
-
checkpoint = torch.load(
|
| 44 |
-
"checkpoints/libero_10_fixed_training_v1/epoch_90.pt",
|
| 45 |
-
map_location="cuda"
|
| 46 |
-
)
|
| 47 |
-
|
| 48 |
-
# Extract model state
|
| 49 |
-
model_state = checkpoint['model_state_dict']
|
| 50 |
-
|
| 51 |
-
# TODO: Add inference code here
|
| 52 |
-
```
|
| 53 |
|
| 54 |
## Performance
|
| 55 |
|
| 56 |
Training run: `libero_10_fixed_training_v1`
|
| 57 |
|
| 58 |
-
*
|
| 59 |
|
| 60 |
## Dataset
|
| 61 |
|
|
@@ -67,11 +48,12 @@ This model was trained on the [GATE-VLAP Datasets](https://huggingface.co/datase
|
|
| 67 |
## Citation
|
| 68 |
|
| 69 |
```bibtex
|
| 70 |
-
@article{
|
| 71 |
-
title={
|
| 72 |
-
author={
|
| 73 |
journal={arXiv preprint arXiv:XXXX.XXXXX},
|
| 74 |
-
|
|
|
|
| 75 |
}
|
| 76 |
```
|
| 77 |
|
|
@@ -83,8 +65,5 @@ This model was trained on the [GATE-VLAP Datasets](https://huggingface.co/datase
|
|
| 83 |
|
| 84 |
- π€ **Dataset**: [gate-institute/GATE-VLAP-datasets](https://huggingface.co/datasets/gate-institute/GATE-VLAP-datasets)
|
| 85 |
- π **Paper**: *Coming soon*
|
| 86 |
-
- π» **Code**: *
|
| 87 |
-
|
| 88 |
-
## License
|
| 89 |
|
| 90 |
-
MIT License
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
library_name: pytorch
|
| 3 |
tags:
|
| 4 |
- robotics
|
|
|
|
| 30 |
- **Dataset**: LIBERO-10 (29 subtasks, 1,354 demonstrations)
|
| 31 |
- **Segmentation**: Semantic action chunking using Gemini Vision API
|
| 32 |
- **Framework**: PyTorch
|
| 33 |
+
- **Checkpoint**: Epoch 90 (best_epoch)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 34 |
|
| 35 |
## Performance
|
| 36 |
|
| 37 |
Training run: `libero_10_fixed_training_v1`
|
| 38 |
|
| 39 |
+
*Overall performance accuracy: 88.8 % task success rate => 5 % better than raw CLIP-RT on LIBERO-LONG*
|
| 40 |
|
| 41 |
## Dataset
|
| 42 |
|
|
|
|
| 48 |
## Citation
|
| 49 |
|
| 50 |
```bibtex
|
| 51 |
+
@article{gateVLAP@SAC2026,
|
| 52 |
+
title={Atomic Action Slicing: Planner-Aligned Options for Generalist VLA Agents},
|
| 53 |
+
author={Stefan Tabakov, Asen Popov, Dimitar Dimitrov, Ensiye Kiyamousavi and Boris Kraychev},
|
| 54 |
journal={arXiv preprint arXiv:XXXX.XXXXX},
|
| 55 |
+
conference={The 41st ACM/SIGAPP Symposium On Applied Computing (SAC2026), track on Intelligent Robotics and Multi-Agent Systems (IRMAS)},
|
| 56 |
+
year={2025}
|
| 57 |
}
|
| 58 |
```
|
| 59 |
|
|
|
|
| 65 |
|
| 66 |
- π€ **Dataset**: [gate-institute/GATE-VLAP-datasets](https://huggingface.co/datasets/gate-institute/GATE-VLAP-datasets)
|
| 67 |
- π **Paper**: *Coming soon*
|
| 68 |
+
- π» **Code**: *Coming soon*
|
|
|
|
|
|
|
| 69 |
|
|
|