|
|
--- |
|
|
license: mit |
|
|
tags: |
|
|
- vision |
|
|
- robotics |
|
|
- drone-navigation |
|
|
- vit |
|
|
--- |
|
|
|
|
|
# autonomous_drone_nav_vision |
|
|
|
|
|
## Overview |
|
|
A Vision Transformer (ViT) fine-tuned for tactical aerial navigation. This model enables Small Unmanned Aircraft Systems (sUAS) to classify environmental obstacles and identify safe landing zones in real-time using downward and forward-facing RGB cameras. |
|
|
|
|
|
|
|
|
|
|
|
## Model Architecture |
|
|
The model utilizes a **Vision Transformer (ViT-Base)** backbone: |
|
|
- **Patch Extraction**: Images are divided into $16 \times 16$ fixed-size patches. |
|
|
- **Position Embeddings**: Learnable spatial embeddings are added to the patch sequence to retain structural context. |
|
|
- **Attention Mechanism**: Global self-attention allows the model to correlate distant visual cues, such as horizon lines and ground markers. |
|
|
|
|
|
## Intended Use |
|
|
- **Obstacle Avoidance**: Integrated into flight control stacks for autonomous "sense and avoid" maneuvers. |
|
|
- **Precision Landing**: Identifying designated markers or flat terrain for autonomous recovery. |
|
|
- **Search and Rescue**: Preliminary screening of aerial footage to identify human-made structures or anomalies. |
|
|
|
|
|
## Limitations |
|
|
- **Low Light**: Performance degrades significantly in nighttime or heavy fog conditions without thermal input. |
|
|
- **Motion Blur**: Rapid yaw movements at high speeds may cause misclassification due to pixel streaking. |
|
|
- **Scale Invariance**: Small objects at extreme altitudes may be missed due to the fixed $224 \times 224$ input resolution. |