File size: 1,517 Bytes
eaf6b10
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5b1a4e4
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
---
license: agpl-3.0
library_name: ultralytics
pipeline_tag: object-detection
tags:
- yolo
- yolo11
- fire-detection
- computer-vision
- realtime
---

# YOLOv11n Fire Detector (ProFSAM)

**Paper:** https://arxiv.org/abs/2510.21782  
**Code:** https://github.com/UEmmanuel5/ProFSAM  
**Weights:** `Fire_best.pt`  

## Intended use
Bounding-box detection of fire to prompt SAM2/MobileSAM/TinySAM in the ProFSAM pipeline.

## Training data
FASDD subset: classes **fire** and **neither_firenorsmoke** only. Total images used: **51,749** (12,550 fire, 39,199 neither_firenorsmoke).

## Training setup (summary)
PyTorch 2.0, CUDA 12.4, main GPU: GTX 1050 Ti (4 GB).  
Ultralytics YOLOv11n initialized then trained 100 epochs.

### Script used
```python
# train_yolo11n.py
import os
os.environ["CUDA_VISIBLE_DEVICES"] = "0"
from ultralytics import YOLO
import torch
torch.backends.cudnn.benchmark = True

model = YOLO("path/to/yolo11n.pt")
train_results = model.train(
    data="path/to/FASDD_CV_Fire/data.yaml",
    epochs=100,
    imgsz=640,
    batch=16,
    optimizer="AdamW",
    lr0=1e-4,
    lrf=0.01,
    dropout=0.15,
    weight_decay=5e-4,
    device=0,
    val=False,
    save=True,
    plots=False
)
````

## Detector metrics (FASDD fire-only subset)

| P     | R     | mAP@0.5 | mAP@0.5:0.95 |
| ----- | ----- | ------- | ------------ |
| 0.799 | 0.697 | 0.797   | 0.520        |


## Test data
If you do not have test images, I placed 4 test images from the khan dataset to be used during your testing phase.