ander2221/visdrone-yolo9-preview

YOLOv9-t fine-tuned on VisDrone2019-DET aerial imagery using LibreYOLO. Ten classes (pedestrian, people, bicycle, car, van, truck, tricycle, awning-tricycle, bus, motor), top-down drone perspective.

Companion use case: LibreYOLO/use-cases/visdrone-finetune.

Training

  • size: t
  • imgsz: 384
  • epochs: 5
  • dataset: VisDrone2019-DET via Voxel51's HuggingFace mirror
  • compute: Apple Metal Performance Shaders (MPS, M-series GPU)

Metrics

{}

Usage β€” Python

from huggingface_hub import hf_hub_download
from libreyolo import LibreYOLO

ckpt = hf_hub_download(repo_id="ander2221/visdrone-yolo9-preview", filename="visdrone.pt")
model = LibreYOLO(ckpt)
result = model("aerial.jpg")
for box, cls, conf in zip(result.boxes.xyxy, result.boxes.cls, result.boxes.conf):
    print(box, ["pedestrian","people","bicycle","car","van","truck","tricycle","awning-tricycle","bus","motor"][int(cls)], float(conf))

Usage β€” ONNX (browser, edge, cross-runtime)

import onnxruntime as ort
from huggingface_hub import hf_hub_download

onnx = hf_hub_download(repo_id="ander2221/visdrone-yolo9-preview", filename="visdrone.onnx")
session = ort.InferenceSession(onnx, providers=["CPUExecutionProvider"])
# Preprocess image to (1, 3, 384, 384) float32 in [0,1] then:
out = session.run(None, {"images": preprocessed})

A live browser demo using this ONNX is at https://libreyolo.github.io/use-cases/visdrone-finetune/demo/ (zero-install, runs locally in Chrome via WebGPU/onnxruntime-web).

Classes (index β†’ name)

idx name
0 pedestrian
1 people
2 bicycle
3 car
4 van
5 truck
6 tricycle
7 awning-tricycle
8 bus
9 motor

License

MIT (the model file). Dataset (VisDrone2019-DET) is governed by its own license terms β€” please review for your use case.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Dataset used to train ander2221/visdrone-yolo9-preview