ML-GOD commited on
Commit
f23732c
·
verified ·
1 Parent(s): 92b11cc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +51 -3
README.md CHANGED
@@ -1,3 +1,51 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: opentau
3
+ tags:
4
+ - robotics
5
+ - vla
6
+ - pi0
7
+ - libero
8
+ - Reinforcement Learning
9
+ - manipulation
10
+ - flow-matching
11
+ - pytorch
12
+ license: apache-2.0
13
+ datasets:
14
+ - physical-intelligence/libero
15
+ repo_url: https://github.com/TensorAuto/OpenTau
16
+ ---
17
+
18
+ # moka_pot_RECAP_R0
19
+
20
+ A **pi0 (π₀) RECAP** Vision-Language-Action (VLA) model, finetuned on the **LIBERO** robotic manipulation benchmark using the **OpenTau** training framework. This model is designed to follow natural language instructions to perform manipulation tasks in a simulated tabletop environment.
21
+ Achieves **~89% success rate** measured over **320 episodes**.
22
+
23
+ **For full documentation, evaluation results, and inference code, please visit the repository:**
24
+ <br>
25
+ 👉 **[https://github.com/TensorAuto/OpenTau](https://github.com/TensorAuto/OpenTau)**
26
+
27
+ ---
28
+
29
+ ## Model Details
30
+
31
+ ### Description
32
+ - **Model Type:** Vision-Language-Action (VLA) Model
33
+ - **Base Architecture:** π₀ (pi0) by Physical Intelligence
34
+ - **Backbone:** PaliGemma-3B (VLM) + Gemma-300M (Action Expert) + RL indicator
35
+ - **Training Data:** Moka Pot Task on LIBERO (Lifelong Robot Learning) Benchmark
36
+ - **Framework:** OpenTau
37
+
38
+ ### Architecture
39
+ The **PI0 RECAP** architecture uses a flow-matching and Reinforcement Learning policy designed for open-world generalization. It combines a Visual Language Model (VLM) for high-level semantic understanding with a smaller "action expert" model that generates continuous joint trajectories (10-step action chunks) via flow matching. It uses RL to learn from good and bad episodes
40
+
41
+ ---
42
+
43
+ ## Training and Evaluation
44
+ The Advantage Indicator (It) was set to True for only 10% of datapoints.
45
+
46
+ ### Dataset
47
+ This model was finetuned on the **Moka Pot task in LIBERO 10** benchmark dataset and autonomous rollouts. It consists of around 29 expert teleoperated episodes and 212 autonomous rollouts under moka_pot_libero_sft policy.
48
+
49
+ ### Results
50
+ For detailed usage instructions, success rates, baseline comparisons, and evaluation protocols, please refer to the [OpenTau GitHub Repository](https://github.com/TensorAuto/OpenTau).
51
+ Achieves **~89% success rate** measured over **320 episodes**.