File size: 4,995 Bytes
f23a73a 98f47c2 f23a73a 98f47c2 5bf1ac4 f923d1a 5bf1ac4 f23a73a 98f47c2 f23a73a 5bf1ac4 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 5bf1ac4 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 5bf1ac4 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 f23a73a 98f47c2 5bf1ac4 98f47c2 5bf1ac4 98f47c2 f23a73a 98f47c2 f23a73a 5bf1ac4 98f47c2 f23a73a 98f47c2 5bf1ac4 98f47c2 f23a73a 98f47c2 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 | ---
license: mit
language:
- en
library_name: pytorch
tags:
- computer-vision
- autonomous-driving
- self-driving
- interfuser
- carla
- object-detection
- trajectory-prediction
datasets:
- PDM-Lite-CARLA
pipeline_tag: object-detection
---
# ๐ InterFuser-Baseer-v1: Autonomous Driving Model
[](https://opensource.org/licenses/MIT)
[](https://pytorch.org/)
[](https://carla.org/)
[](https://huggingface.co/spaces/BaseerAI/Baseer_Server)
> **๐ฎ [Try the Live Demo](https://huggingface.co/spaces/BaseerAI/Baseer_Server)** - Experience the model in action with real-time autonomous driving simulation!
## ๐ Overview
InterFuser-Baseer-v1 is a state-of-the-art transformer-based model for autonomous driving, specifically fine-tuned for the **[Baseer Self-Driving API](https://huggingface.co/spaces/BaseerAI/Baseer_Server)**. This model combines computer vision and deep learning to provide real-time traffic object detection and trajectory planning in simulated driving environments.
### ๐ฏ Key Capabilities
- **Multi-Task Learning**: Simultaneous traffic object detection and waypoint prediction
- **Transformer Architecture**: Advanced attention mechanisms for scene understanding
- **Real-Time Processing**: Optimized for real-time inference in driving scenarios
- **CARLA Integration**: Specifically tuned for CARLA simulation environment
## ๐๏ธ Architecture
### Model Components
| Component | Specification |
|-----------|---------------|
| **Image Backbone** | ResNet-50 (ImageNet pretrained) |
| **LiDAR Backbone** | ResNet-18 (disabled in this version) |
| **Transformer** | 6-layer encoder/decoder, 8 attention heads |
| **Embedding Dimension** | 256 |
| **Prediction Heads** | GRU-based waypoint predictor + Detection head |
### Output Format
- **Traffic Detection**: 20ร20ร7 grid (confidence, position, dimensions, orientation)
- **Waypoint Prediction**: 10 future trajectory points
- **Scene Understanding**: Junction, traffic light, and stop sign detection
## ๐ Quick Start
### Installation
```bash
pip install torch torchvision timm huggingface_hub
```
### Usage Example
```python
import torch
from huggingface_hub import hf_hub_download
# Download model weights
model_path = hf_hub_download(
repo_id="BaseerAI/Interfuser-Baseer-v1",
filename="best_model.pth"
)
# Load model (requires InterFuser class definition)
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model = torch.load(model_path, map_location=device)
model.eval()
# Inference
with torch.no_grad():
outputs = model(input_data)
```
## ๐ Performance
### Training Details
- **Dataset**: PDM-Lite-CARLA (Urban driving scenarios)
- **Training Objective**: Multi-task learning with IoU optimization
- **Framework**: PyTorch
### Key Metrics
- Optimized for traffic detection accuracy
- Enhanced bounding box IoU performance
- Robust waypoint prediction in urban scenarios
## โ ๏ธ Limitations
### Current Constraints
- **Simulation Only**: Trained exclusively on CARLA data
- **Single Camera**: Front-facing camera view only
- **No LiDAR**: Vision-based approach without LiDAR fusion
- **Dataset Scope**: Limited to PDM-Lite-CARLA scenarios
### Recommended Use Cases
- โ
CARLA simulation environments
- โ
Research and development
- โ
Autonomous driving prototyping
- โ Real-world deployment (requires additional training)
## ๐ ๏ธ Integration
This model is designed to work with:
- **[Baseer Self-Driving API](https://huggingface.co/spaces/BaseerAI/Baseer_Server)** - Live demo and API
- **CARLA Simulator**
- **PyTorch Inference Pipeline**
- **Custom Autonomous Driving Systems**
## ๐ Citation
If you use this model in your research, please cite:
```bibtex
@misc{interfuser-baseer-v1,
title={InterFuser-Baseer-v1: Fine-tuned Autonomous Driving Model},
author={BaseerAI},
year={2024},
publisher={Hugging Face},
howpublished={\url{https://huggingface.co/BaseerAI/Interfuser-Baseer-v1}}
}
```
## ๐จโ๐ป Development
**Developed by**: Adam Altawil
**Project Type**: Graduation Project - AI & Autonomous Driving
**Institution**: [Your Institution Name]
## ๐ License
This project is licensed under the MIT License - see the [LICENSE](LICENSE) file for details.
## ๐ค Contributing
Contributions, issues, and feature requests are welcome! Feel free to check the [issues page](../../issues).
## ๐ Support
For questions and support:
- Try the live demo: **[Baseer Server Space](https://huggingface.co/spaces/BaseerAI/Baseer_Server)**
- Create an issue in this repository
- Contact: [Your Contact Information]
---
<div align="center">
<strong>๐ Drive the Future with AI ๐</strong>
</div> |