Add model card and metadata for LEAD (#1)
Browse files- Add model card and metadata for LEAD (7e8ec5346a85dfd70104fb41a302169d4306ccbc)
Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>
README.md
CHANGED
|
@@ -1,3 +1,58 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: mit
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
pipeline_tag: robotics
|
| 4 |
+
tags:
|
| 5 |
+
- autonomous-driving
|
| 6 |
+
- imitation-learning
|
| 7 |
+
- carla
|
| 8 |
+
- transfuser
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
# LEAD: Minimizing Learner–Expert Asymmetry in End-to-End Driving
|
| 12 |
+
|
| 13 |
+
[**Project Page**](https://ln2697.github.io/lead) | [**Paper**](https://huggingface.co/papers/2512.20563) | [**Code**](https://github.com/autonomousvision/lead)
|
| 14 |
+
|
| 15 |
+
Official model weights for **LEAD** and **TransFuser v6 (TFv6)**, an expert-student policy pair for autonomous driving research in the CARLA simulator.
|
| 16 |
+
|
| 17 |
+
LEAD addresses the misalignment between privileged expert demonstrations and sensor-based student observations in imitation learning. By narrowing these asymmetries, the TFv6 student policy achieves state-of-the-art performance on major CARLA closed-loop benchmarks.
|
| 18 |
+
|
| 19 |
+
## Main Features
|
| 20 |
+
|
| 21 |
+
- **Lean pipeline**: Pure PyTorch implementation with minimal dependencies.
|
| 22 |
+
- **Cross-dataset training**: Support for NAVSIM and Waymo datasets, with optional co-training on synthetic CARLA data.
|
| 23 |
+
- **Data-centric infrastructure**: Enforced tensor typing with BearType/JaxTyping and extensive visualizations for debugging.
|
| 24 |
+
- **State-of-the-Art Performance**: TFv6 reaches 95 DS on Bench2Drive and significantly outperforms prior models on Longest6 v2 and Town13.
|
| 25 |
+
|
| 26 |
+
## Evaluation Results (Bench2Drive)
|
| 27 |
+
|
| 28 |
+
| Method | Driving Score (DS) | Success Rate (SR) |
|
| 29 |
+
|---|---|---|
|
| 30 |
+
| TF++ (TFv5) | 84.21 | 67.27 |
|
| 31 |
+
| **TFv6 (Ours)** | **95.28** | **86.80** |
|
| 32 |
+
|
| 33 |
+
## Usage
|
| 34 |
+
|
| 35 |
+
For setup instructions, data collection, and evaluation scripts, please refer to the [official GitHub repository](https://github.com/autonomousvision/lead) and the [full documentation](https://ln2697.github.io/lead/docs).
|
| 36 |
+
|
| 37 |
+
Example evaluation command:
|
| 38 |
+
```bash
|
| 39 |
+
bash scripts/start_carla.sh # Start CARLA server
|
| 40 |
+
bash scripts/eval_bench2drive.sh # Evaluate one Bench2Drive route
|
| 41 |
+
```
|
| 42 |
+
|
| 43 |
+
## Citation
|
| 44 |
+
|
| 45 |
+
If you find this work useful, please cite:
|
| 46 |
+
|
| 47 |
+
```bibtex
|
| 48 |
+
@article{Nguyen2025ARXIV,
|
| 49 |
+
title={LEAD: Minimizing Learner-Expert Asymmetry in End-to-End Driving},
|
| 50 |
+
author={Nguyen, Long and Fauth, Micha and Jaeger, Bernhard and Dauner, Daniel and Igl, Maximilian and Geiger, Andreas and Chitta, Kashyap},
|
| 51 |
+
journal={arXiv preprint arXiv:2512.20563},
|
| 52 |
+
year={2025}
|
| 53 |
+
}
|
| 54 |
+
```
|
| 55 |
+
|
| 56 |
+
## License
|
| 57 |
+
|
| 58 |
+
This project is released under the [MIT License](LICENSE).
|