Add model card and metadata for LEAD
Browse filesHi! I'm Niels from the Hugging Face community science team. I'm opening this PR to improve the model card for LEAD.
This PR adds:
- The `robotics` pipeline tag to improve discoverability.
- Metadata including the MIT license.
- Links to the paper, project page, and official GitHub repository.
- A summary of the TransFuser v6 (TFv6) features and results.
- The BibTeX citation for proper attribution.
Adding this information helps users understand the context of the model checkpoints and how to use them within the broader autonomous driving research ecosystem.
README.md
CHANGED
|
@@ -1,3 +1,58 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: mit
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
pipeline_tag: robotics
|
| 4 |
+
tags:
|
| 5 |
+
- autonomous-driving
|
| 6 |
+
- imitation-learning
|
| 7 |
+
- carla
|
| 8 |
+
- transfuser
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
# LEAD: Minimizing Learner–Expert Asymmetry in End-to-End Driving
|
| 12 |
+
|
| 13 |
+
[**Project Page**](https://ln2697.github.io/lead) | [**Paper**](https://huggingface.co/papers/2512.20563) | [**Code**](https://github.com/autonomousvision/lead)
|
| 14 |
+
|
| 15 |
+
Official model weights for **LEAD** and **TransFuser v6 (TFv6)**, an expert-student policy pair for autonomous driving research in the CARLA simulator.
|
| 16 |
+
|
| 17 |
+
LEAD addresses the misalignment between privileged expert demonstrations and sensor-based student observations in imitation learning. By narrowing these asymmetries, the TFv6 student policy achieves state-of-the-art performance on major CARLA closed-loop benchmarks.
|
| 18 |
+
|
| 19 |
+
## Main Features
|
| 20 |
+
|
| 21 |
+
- **Lean pipeline**: Pure PyTorch implementation with minimal dependencies.
|
| 22 |
+
- **Cross-dataset training**: Support for NAVSIM and Waymo datasets, with optional co-training on synthetic CARLA data.
|
| 23 |
+
- **Data-centric infrastructure**: Enforced tensor typing with BearType/JaxTyping and extensive visualizations for debugging.
|
| 24 |
+
- **State-of-the-Art Performance**: TFv6 reaches 95 DS on Bench2Drive and significantly outperforms prior models on Longest6 v2 and Town13.
|
| 25 |
+
|
| 26 |
+
## Evaluation Results (Bench2Drive)
|
| 27 |
+
|
| 28 |
+
| Method | Driving Score (DS) | Success Rate (SR) |
|
| 29 |
+
|---|---|---|
|
| 30 |
+
| TF++ (TFv5) | 84.21 | 67.27 |
|
| 31 |
+
| **TFv6 (Ours)** | **95.28** | **86.80** |
|
| 32 |
+
|
| 33 |
+
## Usage
|
| 34 |
+
|
| 35 |
+
For setup instructions, data collection, and evaluation scripts, please refer to the [official GitHub repository](https://github.com/autonomousvision/lead) and the [full documentation](https://ln2697.github.io/lead/docs).
|
| 36 |
+
|
| 37 |
+
Example evaluation command:
|
| 38 |
+
```bash
|
| 39 |
+
bash scripts/start_carla.sh # Start CARLA server
|
| 40 |
+
bash scripts/eval_bench2drive.sh # Evaluate one Bench2Drive route
|
| 41 |
+
```
|
| 42 |
+
|
| 43 |
+
## Citation
|
| 44 |
+
|
| 45 |
+
If you find this work useful, please cite:
|
| 46 |
+
|
| 47 |
+
```bibtex
|
| 48 |
+
@article{Nguyen2025ARXIV,
|
| 49 |
+
title={LEAD: Minimizing Learner-Expert Asymmetry in End-to-End Driving},
|
| 50 |
+
author={Nguyen, Long and Fauth, Micha and Jaeger, Bernhard and Dauner, Daniel and Igl, Maximilian and Geiger, Andreas and Chitta, Kashyap},
|
| 51 |
+
journal={arXiv preprint arXiv:2512.20563},
|
| 52 |
+
year={2025}
|
| 53 |
+
}
|
| 54 |
+
```
|
| 55 |
+
|
| 56 |
+
## License
|
| 57 |
+
|
| 58 |
+
This project is released under the [MIT License](LICENSE).
|