asenppopov commited on
Commit
55a2fbc
Β·
verified Β·
1 Parent(s): 66ad39f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -29
README.md CHANGED
@@ -1,5 +1,4 @@
1
  ---
2
- license: mit
3
  library_name: pytorch
4
  tags:
5
  - robotics
@@ -31,31 +30,13 @@ This model is trained for robotic manipulation tasks using vision-language-actio
31
  - **Dataset**: LIBERO-10 (29 subtasks, 1,354 demonstrations)
32
  - **Segmentation**: Semantic action chunking using Gemini Vision API
33
  - **Framework**: PyTorch
34
- - **Checkpoint**: Epoch 90
35
-
36
- ## Usage
37
-
38
- ```python
39
- import torch
40
- from pathlib import Path
41
-
42
- # Load checkpoint
43
- checkpoint = torch.load(
44
- "checkpoints/libero_10_fixed_training_v1/epoch_90.pt",
45
- map_location="cuda"
46
- )
47
-
48
- # Extract model state
49
- model_state = checkpoint['model_state_dict']
50
-
51
- # TODO: Add inference code here
52
- ```
53
 
54
  ## Performance
55
 
56
  Training run: `libero_10_fixed_training_v1`
57
 
58
- *Add your metrics here after evaluation*
59
 
60
  ## Dataset
61
 
@@ -67,11 +48,12 @@ This model was trained on the [GATE-VLAP Datasets](https://huggingface.co/datase
67
  ## Citation
68
 
69
  ```bibtex
70
- @article{gateVLAP2024,
71
- title={GATE-VLAP: Grounded Action Trajectory Embeddings with Vision-Language Action Planning},
72
- author={[Your Name]},
73
  journal={arXiv preprint arXiv:XXXX.XXXXX},
74
- year={2024}
 
75
  }
76
  ```
77
 
@@ -83,8 +65,5 @@ This model was trained on the [GATE-VLAP Datasets](https://huggingface.co/datase
83
 
84
  - πŸ€— **Dataset**: [gate-institute/GATE-VLAP-datasets](https://huggingface.co/datasets/gate-institute/GATE-VLAP-datasets)
85
  - πŸ“„ **Paper**: *Coming soon*
86
- - πŸ’» **Code**: *Add your GitHub repo here*
87
-
88
- ## License
89
 
90
- MIT License
 
1
  ---
 
2
  library_name: pytorch
3
  tags:
4
  - robotics
 
30
  - **Dataset**: LIBERO-10 (29 subtasks, 1,354 demonstrations)
31
  - **Segmentation**: Semantic action chunking using Gemini Vision API
32
  - **Framework**: PyTorch
33
+ - **Checkpoint**: Epoch 90 (best_epoch)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
 
35
  ## Performance
36
 
37
  Training run: `libero_10_fixed_training_v1`
38
 
39
+ *Overall performance accuracy: 88.8 % task success rate => 5 % better than raw CLIP-RT on LIBERO-LONG*
40
 
41
  ## Dataset
42
 
 
48
  ## Citation
49
 
50
  ```bibtex
51
+ @article{gateVLAP@SAC2026,
52
+ title={Atomic Action Slicing: Planner-Aligned Options for Generalist VLA Agents},
53
+ author={Stefan Tabakov, Asen Popov, Dimitar Dimitrov, Ensiye Kiyamousavi and Boris Kraychev},
54
  journal={arXiv preprint arXiv:XXXX.XXXXX},
55
+ conference={The 41st ACM/SIGAPP Symposium On Applied Computing (SAC2026), track on Intelligent Robotics and Multi-Agent Systems (IRMAS)},
56
+ year={2025}
57
  }
58
  ```
59
 
 
65
 
66
  - πŸ€— **Dataset**: [gate-institute/GATE-VLAP-datasets](https://huggingface.co/datasets/gate-institute/GATE-VLAP-datasets)
67
  - πŸ“„ **Paper**: *Coming soon*
68
+ - πŸ’» **Code**: *Coming soon*
 
 
69