LeTau commited on
Commit
1857875
Β·
verified Β·
1 Parent(s): 324868f

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +7 -7
README.md CHANGED
@@ -75,7 +75,7 @@ Evaluated on 10 test samples from 1000 synthetic demonstrations:
75
  | Metric | Value | Notes |
76
  |--------|-------|-------|
77
  | Position Error | **8.60cm** | Suitable for ~5cm cube picking |
78
- | Gripper Accuracy | **70%** | Reliable grasp planning |
79
  | Overall MAE | **0.1217** | Across all 8 action dimensions |
80
  | Quaternion Error | 19.36Β° | Best for top-down grasps |
81
 
@@ -97,7 +97,7 @@ import torch
97
 
98
  # Load model
99
  device = 'cuda' if torch.cuda.is_available() else 'cpu'
100
- checkpoint = torch.load('pytorch_model.bin', map_location=device)
101
 
102
  vlm_encoder = VLM_Encoder().to(device)
103
  policy = ImprovedFlowMatchingPolicy(action_dim=8, context_dim=1024, hidden_dim=512).to(device)
@@ -131,10 +131,10 @@ python vla_flow_matching.py --mode replay --checkpoint vla_checkpoint_best.pt
131
 
132
  ```
133
  β”œβ”€β”€ vla_flow_matching.py # Complete implementation (~1000 lines)
134
- β”œβ”€β”€ pytorch_model.bin # Trained weights (~20MB)
135
- β”œβ”€β”€ demo_data.pkl # Training data (1000 demos)
136
- β”œβ”€β”€ replay_results.png # Evaluation visualization
137
- └── README.md # This file
138
  ```
139
 
140
  ## 🎯 What You'll Learn
@@ -202,7 +202,7 @@ The synthetic environment generates pick-and-place demonstrations:
202
  ```bash
203
  # Collect 10-50 real demonstrations, then:
204
  python vla_flow_matching.py --mode finetune \
205
- --checkpoint pytorch_model.bin \
206
  --data_path real_robot_demos.pkl \
207
  --epochs 30 --lr 1e-5
208
  ```
 
75
  | Metric | Value | Notes |
76
  |--------|-------|-------|
77
  | Position Error | **8.60cm** | Suitable for ~5cm cube picking |
78
+ | Gripper Accuracy | **75%** | Reliable grasp planning |
79
  | Overall MAE | **0.1217** | Across all 8 action dimensions |
80
  | Quaternion Error | 19.36Β° | Best for top-down grasps |
81
 
 
97
 
98
  # Load model
99
  device = 'cuda' if torch.cuda.is_available() else 'cpu'
100
+ checkpoint = torch.load('vla_checkpoint_best.pt', map_location=device)
101
 
102
  vlm_encoder = VLM_Encoder().to(device)
103
  policy = ImprovedFlowMatchingPolicy(action_dim=8, context_dim=1024, hidden_dim=512).to(device)
 
131
 
132
  ```
133
  β”œβ”€β”€ vla_flow_matching.py # Complete implementation (~1000 lines)
134
+ β”œβ”€β”€ vla_checkpoint_best.pt # Trained weights (~20MB)
135
+ β”œβ”€β”€ demo_data.pkl # Training data (1000 demos)
136
+ β”œβ”€β”€ replay_results.png # Evaluation visualization
137
+ └── README.md # This file
138
  ```
139
 
140
  ## 🎯 What You'll Learn
 
202
  ```bash
203
  # Collect 10-50 real demonstrations, then:
204
  python vla_flow_matching.py --mode finetune \
205
+ --checkpoint vla_checkpoint_best.pt \
206
  --data_path real_robot_demos.pkl \
207
  --epochs 30 --lr 1e-5
208
  ```