File size: 4,995 Bytes
f23a73a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
98f47c2
f23a73a
98f47c2
 
 
5bf1ac4
f923d1a
5bf1ac4
f23a73a
98f47c2
f23a73a
5bf1ac4
f23a73a
98f47c2
f23a73a
98f47c2
 
 
 
f23a73a
98f47c2
f23a73a
98f47c2
f23a73a
98f47c2
 
 
 
 
 
 
f23a73a
98f47c2
f23a73a
98f47c2
 
 
f23a73a
98f47c2
 
 
f23a73a
 
 
98f47c2
f23a73a
98f47c2
f23a73a
98f47c2
f23a73a
98f47c2
f23a73a
98f47c2
 
5bf1ac4
98f47c2
 
 
 
f23a73a
98f47c2
 
 
 
 
 
 
 
 
 
 
f23a73a
98f47c2
 
 
f23a73a
98f47c2
f23a73a
98f47c2
 
 
f23a73a
98f47c2
f23a73a
98f47c2
f23a73a
98f47c2
 
 
 
f23a73a
98f47c2
f23a73a
98f47c2
 
 
 
f23a73a
98f47c2
f23a73a
98f47c2
f23a73a
5bf1ac4
98f47c2
 
 
f23a73a
98f47c2
f23a73a
98f47c2
f23a73a
98f47c2
 
 
5bf1ac4
98f47c2
 
5bf1ac4
98f47c2
 
f23a73a
98f47c2
f23a73a
5bf1ac4
98f47c2
 
f23a73a
98f47c2
 
 
 
 
 
 
 
 
 
 
5bf1ac4
98f47c2
 
 
 
f23a73a
98f47c2
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
---
license: mit
language:
  - en
library_name: pytorch
tags:
  - computer-vision
  - autonomous-driving
  - self-driving
  - interfuser
  - carla
  - object-detection
  - trajectory-prediction
datasets:
  - PDM-Lite-CARLA
pipeline_tag: object-detection
---

# ๐Ÿš— InterFuser-Baseer-v1: Autonomous Driving Model

[![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT)
[![PyTorch](https://img.shields.io/badge/PyTorch-EE4C2C?style=flat&logo=pytorch&logoColor=white)](https://pytorch.org/)
[![CARLA](https://img.shields.io/badge/CARLA-Simulator-blue)](https://carla.org/)
[![Demo](https://img.shields.io/badge/๐Ÿš€-Live%20Demo-brightgreen)](https://huggingface.co/spaces/BaseerAI/Baseer_Server)

> **๐ŸŽฎ [Try the Live Demo](https://huggingface.co/spaces/BaseerAI/Baseer_Server)** - Experience the model in action with real-time autonomous driving simulation!

## ๐Ÿ“– Overview

InterFuser-Baseer-v1 is a state-of-the-art transformer-based model for autonomous driving, specifically fine-tuned for the **[Baseer Self-Driving API](https://huggingface.co/spaces/BaseerAI/Baseer_Server)**. This model combines computer vision and deep learning to provide real-time traffic object detection and trajectory planning in simulated driving environments.

### ๐ŸŽฏ Key Capabilities

- **Multi-Task Learning**: Simultaneous traffic object detection and waypoint prediction
- **Transformer Architecture**: Advanced attention mechanisms for scene understanding
- **Real-Time Processing**: Optimized for real-time inference in driving scenarios
- **CARLA Integration**: Specifically tuned for CARLA simulation environment

## ๐Ÿ—๏ธ Architecture

### Model Components

| Component | Specification |
|-----------|---------------|
| **Image Backbone** | ResNet-50 (ImageNet pretrained) |
| **LiDAR Backbone** | ResNet-18 (disabled in this version) |
| **Transformer** | 6-layer encoder/decoder, 8 attention heads |
| **Embedding Dimension** | 256 |
| **Prediction Heads** | GRU-based waypoint predictor + Detection head |

### Output Format

- **Traffic Detection**: 20ร—20ร—7 grid (confidence, position, dimensions, orientation)
- **Waypoint Prediction**: 10 future trajectory points
- **Scene Understanding**: Junction, traffic light, and stop sign detection

## ๐Ÿš€ Quick Start

### Installation

```bash
pip install torch torchvision timm huggingface_hub
```

### Usage Example

```python
import torch
from huggingface_hub import hf_hub_download

# Download model weights
model_path = hf_hub_download(
    repo_id="BaseerAI/Interfuser-Baseer-v1",
    filename="best_model.pth"
)

# Load model (requires InterFuser class definition)
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model = torch.load(model_path, map_location=device)
model.eval()

# Inference
with torch.no_grad():
    outputs = model(input_data)
```

## ๐Ÿ“Š Performance

### Training Details

- **Dataset**: PDM-Lite-CARLA (Urban driving scenarios)
- **Training Objective**: Multi-task learning with IoU optimization
- **Framework**: PyTorch

### Key Metrics

- Optimized for traffic detection accuracy
- Enhanced bounding box IoU performance
- Robust waypoint prediction in urban scenarios

## โš ๏ธ Limitations

### Current Constraints

- **Simulation Only**: Trained exclusively on CARLA data
- **Single Camera**: Front-facing camera view only
- **No LiDAR**: Vision-based approach without LiDAR fusion
- **Dataset Scope**: Limited to PDM-Lite-CARLA scenarios

### Recommended Use Cases

- โœ… CARLA simulation environments
- โœ… Research and development
- โœ… Autonomous driving prototyping
- โŒ Real-world deployment (requires additional training)

## ๐Ÿ› ๏ธ Integration

This model is designed to work with:

- **[Baseer Self-Driving API](https://huggingface.co/spaces/BaseerAI/Baseer_Server)** - Live demo and API
- **CARLA Simulator**
- **PyTorch Inference Pipeline**
- **Custom Autonomous Driving Systems**

## ๐Ÿ“š Citation

If you use this model in your research, please cite:

```bibtex
@misc{interfuser-baseer-v1,
  title={InterFuser-Baseer-v1: Fine-tuned Autonomous Driving Model},
  author={BaseerAI},
  year={2024},
  publisher={Hugging Face},
  howpublished={\url{https://huggingface.co/BaseerAI/Interfuser-Baseer-v1}}
}
```

## ๐Ÿ‘จโ€๐Ÿ’ป Development

**Developed by**: Adam Altawil
**Project Type**: Graduation Project - AI & Autonomous Driving  
**Institution**: [Your Institution Name]

## ๐Ÿ“„ License

This project is licensed under the MIT License - see the [LICENSE](LICENSE) file for details.

## ๐Ÿค Contributing

Contributions, issues, and feature requests are welcome! Feel free to check the [issues page](../../issues).

## ๐Ÿ“ž Support

For questions and support:
- Try the live demo: **[Baseer Server Space](https://huggingface.co/spaces/BaseerAI/Baseer_Server)**
- Create an issue in this repository
- Contact: [Your Contact Information]

---

<div align="center">
  <strong>๐Ÿš— Drive the Future with AI ๐Ÿš—</strong>
</div>