Add metadata and improve model card

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +55 -3
README.md CHANGED
@@ -1,19 +1,71 @@
1
  ---
2
  license: apache-2.0
 
 
3
  tags:
4
- - robotics,
5
  - lerobot
6
  ---
7
 
 
 
8
  DiT-EncDec base checkpoint from "CLARE: Continual Learning for Vision-Language-Action Models via Autonomous Adapter Routing and Expansion", pretrained on LIBERO-90.
9
 
10
- **BibTeX:**
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
 
12
  ```bibtex
13
  @article{romer2026clare,
14
  title={CLARE: Continual Learning for Vision-Language-Action Models via Autonomous Adapter Routing and Expansion},
15
  author={Ralf R{\"o}mer and Yi Zhang and Angela P. Schoellig},
16
- journal={arXiv preprint arXiv:2601.09512} ,
17
  year={2026}
18
  }
19
  ```
 
1
  ---
2
  license: apache-2.0
3
+ library_name: lerobot
4
+ pipeline_tag: robotics
5
  tags:
6
+ - robotics
7
  - lerobot
8
  ---
9
 
10
+ # CLARE: Continual Learning for Vision-Language-Action Models via Autonomous Adapter Routing and Expansion
11
+
12
  DiT-EncDec base checkpoint from "CLARE: Continual Learning for Vision-Language-Action Models via Autonomous Adapter Routing and Expansion", pretrained on LIBERO-90.
13
 
14
+ - **Paper:** [CLARE: Continual Learning for Vision-Language-Action Models via Autonomous Adapter Routing and Expansion](https://huggingface.co/papers/2601.09512)
15
+ - **Project Page:** [tum-lsy.github.io/clare/](https://tum-lsy.github.io/clare/)
16
+ - **Repository:** [utiasDSL/clare](https://github.com/utiasDSL/clare)
17
+
18
+ ## Description
19
+
20
+ CLARE is a general, parameter-efficient framework for exemplar-free continual learning with Vision-Language-Action (VLA) models. It introduces lightweight modular adapters into selected feedforward layers and autonomously expands the model only where necessary when learning a new task, guided by layer-wise feature similarity. During deployment, an autoencoder-based routing mechanism dynamically activates the most relevant adapters without requiring task labels.
21
+
22
+ This specific repository contains the **DiT-EncDec** base checkpoint used for pretraining on the **LIBERO-90** benchmark.
23
+
24
+ ## Usage
25
+
26
+ To use this checkpoint for training on the LIBERO-10 benchmark using the CLARE framework, you can use the following command from the [official repository](https://github.com/utiasDSL/clare):
27
+
28
+ ```bash
29
+ python ./lerobot_lsy/src/lerobot/scripts/clare.py \
30
+ --seed=1000 \
31
+ --job_name=clare_libero_10_task_0 \
32
+ --output_dir=./outputs/clare_libero_10_task_0 \
33
+ --dataset.repo_id=continuallearning/libero_10_image_task_0 \
34
+ --policy.path=continuallearning/dit_mt_libero_90_pretrain \
35
+ --policy.push_to_hub=false \
36
+ --batch_size=32 \
37
+ --num_workers=16 \
38
+ --steps=20000 \
39
+ --env.type=libero \
40
+ --env.task=Libero_10_Task_0 \
41
+ --eval.batch_size=20 \
42
+ --eval.n_episodes=100 \
43
+ --eval.max_episodes_rendered=100 \
44
+ --eval_freq=200000 \
45
+ --save_freq=20000 \
46
+ --log_freq=100 \
47
+ --peft_cfg_path=./peft_lsy/config \
48
+ --expand_threshold=10.00 \
49
+ --detect_distribution_shift_steps=200 \
50
+ --detect_distribution_shift_batch_size=32 \
51
+ --detect_distribution_shift_num_workers=16 \
52
+ --detect_distribution_shift_log_freq=10 \
53
+ --train_discriminators_steps=2000 \
54
+ --train_discriminators_batch_size=32 \
55
+ --train_discriminators_num_workers=16 \
56
+ --train_discriminators_log_freq=50 \
57
+ --train_discriminators_eval_freq=2000 \
58
+ --train_discriminators_save_freq=2000 \
59
+ --wandb.enable=true
60
+ ```
61
+
62
+ ## BibTeX
63
 
64
  ```bibtex
65
  @article{romer2026clare,
66
  title={CLARE: Continual Learning for Vision-Language-Action Models via Autonomous Adapter Routing and Expansion},
67
  author={Ralf R{\"o}mer and Yi Zhang and Angela P. Schoellig},
68
+ journal={arXiv preprint arXiv:2601.09512},
69
  year={2026}
70
  }
71
  ```