Upload folder using huggingface_hub
Browse files
README.md
CHANGED
|
@@ -75,7 +75,7 @@ Evaluated on 10 test samples from 1000 synthetic demonstrations:
|
|
| 75 |
| Metric | Value | Notes |
|
| 76 |
|--------|-------|-------|
|
| 77 |
| Position Error | **8.60cm** | Suitable for ~5cm cube picking |
|
| 78 |
-
| Gripper Accuracy | **
|
| 79 |
| Overall MAE | **0.1217** | Across all 8 action dimensions |
|
| 80 |
| Quaternion Error | 19.36Β° | Best for top-down grasps |
|
| 81 |
|
|
@@ -97,7 +97,7 @@ import torch
|
|
| 97 |
|
| 98 |
# Load model
|
| 99 |
device = 'cuda' if torch.cuda.is_available() else 'cpu'
|
| 100 |
-
checkpoint = torch.load('
|
| 101 |
|
| 102 |
vlm_encoder = VLM_Encoder().to(device)
|
| 103 |
policy = ImprovedFlowMatchingPolicy(action_dim=8, context_dim=1024, hidden_dim=512).to(device)
|
|
@@ -131,10 +131,10 @@ python vla_flow_matching.py --mode replay --checkpoint vla_checkpoint_best.pt
|
|
| 131 |
|
| 132 |
```
|
| 133 |
βββ vla_flow_matching.py # Complete implementation (~1000 lines)
|
| 134 |
-
βββ
|
| 135 |
-
βββ demo_data.pkl
|
| 136 |
-
βββ replay_results.png
|
| 137 |
-
βββ README.md
|
| 138 |
```
|
| 139 |
|
| 140 |
## π― What You'll Learn
|
|
@@ -202,7 +202,7 @@ The synthetic environment generates pick-and-place demonstrations:
|
|
| 202 |
```bash
|
| 203 |
# Collect 10-50 real demonstrations, then:
|
| 204 |
python vla_flow_matching.py --mode finetune \
|
| 205 |
-
--checkpoint
|
| 206 |
--data_path real_robot_demos.pkl \
|
| 207 |
--epochs 30 --lr 1e-5
|
| 208 |
```
|
|
|
|
| 75 |
| Metric | Value | Notes |
|
| 76 |
|--------|-------|-------|
|
| 77 |
| Position Error | **8.60cm** | Suitable for ~5cm cube picking |
|
| 78 |
+
| Gripper Accuracy | **75%** | Reliable grasp planning |
|
| 79 |
| Overall MAE | **0.1217** | Across all 8 action dimensions |
|
| 80 |
| Quaternion Error | 19.36Β° | Best for top-down grasps |
|
| 81 |
|
|
|
|
| 97 |
|
| 98 |
# Load model
|
| 99 |
device = 'cuda' if torch.cuda.is_available() else 'cpu'
|
| 100 |
+
checkpoint = torch.load('vla_checkpoint_best.pt', map_location=device)
|
| 101 |
|
| 102 |
vlm_encoder = VLM_Encoder().to(device)
|
| 103 |
policy = ImprovedFlowMatchingPolicy(action_dim=8, context_dim=1024, hidden_dim=512).to(device)
|
|
|
|
| 131 |
|
| 132 |
```
|
| 133 |
βββ vla_flow_matching.py # Complete implementation (~1000 lines)
|
| 134 |
+
βββ vla_checkpoint_best.pt # Trained weights (~20MB)
|
| 135 |
+
βββ demo_data.pkl # Training data (1000 demos)
|
| 136 |
+
βββ replay_results.png # Evaluation visualization
|
| 137 |
+
βββ README.md # This file
|
| 138 |
```
|
| 139 |
|
| 140 |
## π― What You'll Learn
|
|
|
|
| 202 |
```bash
|
| 203 |
# Collect 10-50 real demonstrations, then:
|
| 204 |
python vla_flow_matching.py --mode finetune \
|
| 205 |
+
--checkpoint vla_checkpoint_best.pt \
|
| 206 |
--data_path real_robot_demos.pkl \
|
| 207 |
--epochs 30 --lr 1e-5
|
| 208 |
```
|