DAVE-2 End-to-End Driving Model
Implementation of NVIDIA's DAVE-2 architecture trained on the Udacity self-driving car simulator for the bachelor's thesis: Dual-Axis Testing of Visual Robustness and Topological Generalization in Vision-based End-to-End Driving Models.
Model Description
DAVE-2 is the original end-to-end driving architecture proposed by NVIDIA in 2016. It learns to map raw camera images directly to steering and throttle commands through imitation learning.
Architecture
Input: RGB Image (66 ร 200 ร 3)
โ
Conv2D(24, 5ร5, stride=2) + ELU
Conv2D(36, 5ร5, stride=2) + ELU
Conv2D(48, 5ร5, stride=2) + ELU
Conv2D(64, 3ร3) + ELU
Conv2D(64, 3ร3) + ELU
โ
Flatten
โ
Dense(1164) + ELU
Dense(100) + ELU
Dense(50) + ELU
Dense(10) + ELU
โ
Output: [steering, throttle]
Checkpoints
| Map | Checkpoint |
|---|---|
| GenRoads | genroads_20251028-145557/ |
| Jungle | jungle_20251209-175046/ |
Files per Checkpoint
best_model.h5: Keras model weightsmeta.json: Training configuration and hyperparametershistory.csv: Training/validation metrics per epochloss_curve.png: Visualization of training progress
Citation
@thesis{igenbergs2026dualaxis,
title={Dual-Axis Testing of Visual Robustness and Topological Generalization in Vision-based End-to-End Driving Models},
author={Igenbergs, Maxim},
school={Technical University of Munich},
year={2026},
type={Bachelor's Thesis}
}
Related
- Downloads last month
- -
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support