Robotics

Improve model card: Add pipeline tag, paper, project page, code links, and sample usage

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +112 -2
README.md CHANGED
@@ -1,10 +1,11 @@
1
  ---
2
- license: apache-2.0
3
  datasets:
4
  - behavior-1k/2025-challenge-demos
5
  - IliaLarchenko/behavior_224_rgb
 
6
  tags:
7
  - robotics
 
8
  ---
9
 
10
  This is an intermediate checkpoint that we used in our [1st place solution of the 2025 BEHAVIOR Challenge](https://github.com/IliaLarchenko/behavior-1k-solution).
@@ -13,10 +14,119 @@ This checkpoint is obtained by training the policy on 50 tasks simultaneously fo
13
 
14
  It is not part of our [final submission](https://huggingface.co/IliaLarchenko/behavior_submission). Also, we didn't run the whole evaluation of this checkpoint, but we would expect it to achieve a 15-20% q-score.
15
 
16
- Our [tech report](https:///arxiv.org/abs/2512.06951)
 
 
 
17
 
18
  The [final submission checkpoints](https://huggingface.co/IliaLarchenko/behavior_submission)
19
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  ## Citation
22
 
 
1
  ---
 
2
  datasets:
3
  - behavior-1k/2025-challenge-demos
4
  - IliaLarchenko/behavior_224_rgb
5
+ license: apache-2.0
6
  tags:
7
  - robotics
8
+ pipeline_tag: robotics
9
  ---
10
 
11
  This is an intermediate checkpoint that we used in our [1st place solution of the 2025 BEHAVIOR Challenge](https://github.com/IliaLarchenko/behavior-1k-solution).
 
14
 
15
  It is not part of our [final submission](https://huggingface.co/IliaLarchenko/behavior_submission). Also, we didn't run the whole evaluation of this checkpoint, but we would expect it to achieve a 15-20% q-score.
16
 
17
+ Paper: [Task adaptation of Vision-Language-Action model: 1st Place Solution for the 2025 BEHAVIOR Challenge](https://huggingface.co/papers/2512.06951)
18
+ Project page: https://behavior.stanford.edu/challenge/
19
+ Code/GitHub Repository: [IliaLarchenko/behavior-1k-solution](https://github.com/IliaLarchenko/behavior-1k-solution)
20
+ arXiv: [2512.06951](https://arxiv.org/abs/2512.06951)
21
 
22
  The [final submission checkpoints](https://huggingface.co/IliaLarchenko/behavior_submission)
23
 
24
+ ## Sample Usage
25
+
26
+ This section provides a quick overview of how to get started with the model, adapted from the [GitHub repository](https://github.com/IliaLarchenko/behavior-1k-solution).
27
+
28
+ ### Installation
29
+
30
+ ```bash
31
+ # Clone with submodules (includes openpi and BEHAVIOR-1K)
32
+ git clone --recurse-submodules https://github.com/ilialarchenko/behavior-1k-solution.git
33
+ cd behavior-1k-solution
34
+
35
+ # Run setup script (installs uv, dependencies, and sets up environment)
36
+ bash setup_remote.sh
37
+ ```
38
+
39
+ ### Dataset Preparation
40
+
41
+ Download the official BEHAVIOR-1K dataset from HuggingFace:
42
+
43
+ ```bash
44
+ # Login to HuggingFace (need to avoid request rate limit errors)
45
+ uv run huggingface-cli login
46
+
47
+ # Download the full dataset (~2TB)
48
+ uv run python - <<'PY'
49
+ from huggingface_hub import snapshot_download
50
+ snapshot_download(
51
+ repo_id="behavior-1k/2025-challenge-demos",
52
+ repo_type="dataset",
53
+ local_dir="./data/behavior_dataset",
54
+ local_dir_use_symlinks=False
55
+ )
56
+ PY
57
+ ```
58
+
59
+ **Alternative**: Use the resized RGB-only dataset (224×224, ~260GB) for faster training:
60
+ ```bash
61
+ uv run python - <<'PY'
62
+ from huggingface_hub import snapshot_download
63
+ snapshot_download(
64
+ repo_id="IliaLarchenko/behavior_224_rgb",
65
+ repo_type="dataset",
66
+ local_dir="./data/behavior_224_rgb",
67
+ local_dir_use_symlinks=False
68
+ )
69
+ PY
70
+ ```
71
+
72
+ ### Pre-training Setup
73
+
74
+ Compute dataset statistics and train FAST tokenizer:
75
+
76
+ ```bash
77
+ # Compute normalization statistics with correlation matrix
78
+ uv run scripts/compute_norm_stats.py --config-name pi_behavior_b1k_fast --correlation
79
+
80
+ # Train FAST tokenizer for action discretization
81
+ uv run scripts/train_fast_tokenizer.py \
82
+ --config-name pi_behavior_b1k_fast \
83
+ --encoded-dims="0:6,7:23" \
84
+ --vocab-size=1024
85
+ ```
86
+
87
+ ### Training
88
+
89
+ **Single GPU Training**:
90
+ ```bash
91
+ uv run scripts/train.py pi_behavior_b1k_fast \
92
+ --batch_size=16 \
93
+ --num_train_steps=200000 \
94
+ --save_interval=2000 \
95
+ --keep_period=10000 \
96
+ --log_interval=100
97
+ ```
98
+
99
+ **Multi-GPU Training**:
100
+ ```bash
101
+ uv run scripts/train.py pi_behavior_b1k_fast \
102
+ --batch_size=2048 \
103
+ --num_train_steps=200000 \
104
+ --fsdp_devices=8 \
105
+ --save_interval=250 \
106
+ --keep_period=4000 \
107
+ --log_interval=25
108
+ ```
109
+
110
+ ### Evaluation
111
+
112
+ Start the policy server:
113
+
114
+ ```bash
115
+ uv run scripts/serve_b1k.py policy:checkpoint \
116
+ --policy.config pi_behavior_b1k_fast \
117
+ --policy.dir /path/to/checkpoint
118
+ ```
119
+
120
+ In a separate terminal, [run evaluation](https://behavior.stanford.edu/challenge/baselines.html) (requires BEHAVIOR-1K environment):
121
+
122
+ ```bash
123
+ python BEHAVIOR-1K/omnigibson/learning/eval.py \
124
+ log_path=./eval_logs \
125
+ policy=websocket \
126
+ task.name=make_microwave_popcorn \
127
+ model.host=localhost \
128
+ eval_instance_ids="[0,1,2,3]"
129
+ ```
130
 
131
  ## Citation
132