ultralytics
Eval Results (legacy)

Ultralytics YOLO26 is a cutting-edge, state-of-the-art (SOTA) model that builds upon the success of previous Ultralytics YOLO versions and introduces new features and improvements to further boost performance and flexibility. Ultralytics YOLO26 is designed to be fast, accurate, and easy to use, making it an excellent choice for a wide range of object detection and tracking, instance segmentation, image classification and pose estimation tasks.

We hope that the resources here will help you get the most out of Ultralytics YOLO. Please browse the Ultralytics Docs for details, raise an issue on GitHub for support, questions, or discussions, become a member of the Ultralytics Discord, Reddit and Forums!

To request an Enterprise License please complete the form at Ultralytics Licensing.

YOLO26 performance plots
Ultralytics GitHub Ultralytics LinkedIn Ultralytics Twitter Ultralytics YouTube Ultralytics TikTok Ultralytics BiliBili Ultralytics Discord

Documentation

See below for a quickstart install and usage examples, and see our Docs for full documentation on training, validation, prediction and deployment.

Install

Pip install the ultralytics package including all requirements in a Python>=3.8 environment with PyTorch>=1.8.

PyPI - Version Downloads PyPI - Python Version
pip install ultralytics

For alternative installation methods including Conda, Docker, and Git, please refer to the Quickstart Guide.

Conda Version Docker Image Version Ultralytics Docker Pulls
Usage

CLI

Ultralytics YOLO may be used directly in the Command Line Interface (CLI) with a yolo command:

yolo predict model=yolo26n.pt source='https://ultralytics.com/images/bus.jpg'

yolo can be used for a variety of tasks and modes and accepts additional arguments, i.e. imgsz=640. See the Ultralytics YOLO CLI Docs for examples.

Python

Ultralytics YOLO may also be used directly in a Python environment, and accepts the same arguments as in the CLI example above:

from ultralytics import YOLO
# Load a model
model = YOLO("yolo26n.pt")
# Train the model
train_results = model.train(
    data="coco8.yaml",  # path to dataset YAML
    epochs=100,  # number of training epochs
    imgsz=640,  # training image size
    device="cpu",  # device to run on, i.e. device=0 or device=0,1,2,3 or device=cpu
)
# Evaluate model performance on the validation set
metrics = model.val()
# Perform object detection on an image
results = model("path/to/image.jpg")
results[0].show()
# Export the model to ONNX format
path = model.export(format="onnx")  # return path to exported model

See Ultralytics YOLO Python Docs for more examples.

Models

Ultralytics YOLO26 Detect, Segment and Pose models pretrained on the COCO dataset are available here, as well as Ultralytics YOLO26 Classify models pretrained on the ImageNet dataset. Track mode is available for all Detect, Segment and Pose models.

Ultralytics YOLO supported tasks

All Models download automatically from the latest Ultralytics release on first use.

Detection (COCO)

See Detection Docs for usage examples with these models trained on COCO, which include 80 pre-trained classes.

Model size
(pixels)
mAPval
50-95
mAPval
50-95(e2e)
Speed
CPU ONNX
(ms)
Speed
T4 TensorRT10
(ms)
params
(M)
FLOPs
(B)
YOLO26n 640 40.9 40.1 38.9 ± 0.7 1.7 ± 0.0 2.4 5.4
YOLO26s 640 48.6 47.8 87.2 ± 0.9 2.5 ± 0.0 9.5 20.7
YOLO26m 640 53.1 52.5 220.0 ± 1.4 4.7 ± 0.1 20.4 68.2
YOLO26l 640 55.0 54.4 286.2 ± 2.0 6.2 ± 0.2 24.8 86.4
YOLO26x 640 57.5 56.9 525.8 ± 4.0 11.8 ± 0.2 55.7 193.9
  • mAPval values are for single-model single-scale on COCO val2017 dataset.
    Reproduce by yolo val detect data=coco.yaml device=0
  • Speed averaged over COCO val images using an Amazon EC2 P4d instance.
    Reproduce by yolo val detect data=coco.yaml batch=1 device=0|cpu
Segmentation (COCO)

See Segmentation Docs for usage examples with these models trained on COCO-Seg, which include 80 pre-trained classes.

Model size
(pixels)
mAPbox
50-95(e2e)
mAPmask
50-95(e2e)
Speed
CPU ONNX
(ms)
Speed
T4 TensorRT10
(ms)
params
(M)
FLOPs
(B)
YOLO26n-seg 640 39.6 33.9 53.3 ± 0.5 2.1 ± 0.0 2.7 9.1
YOLO26s-seg 640 47.3 40.0 118.4 ± 0.9 3.3 ± 0.0 10.4 34.2
YOLO26m-seg 640 52.5 44.1 328.2 ± 2.4 6.7 ± 0.1 23.6 121.5
YOLO26l-seg 640 54.4 45.5 387.0 ± 3.7 8.0 ± 0.1 28.0 139.8
YOLO26x-seg 640 56.5 47.0 787.0 ± 6.8 16.4 ± 0.1 62.8 313.5
  • mAPval values are for single-model single-scale on COCO val2017 dataset.
    Reproduce by yolo val segment data=coco-seg.yaml device=0
  • Speed averaged over COCO val images using an Amazon EC2 P4d instance.
    Reproduce by yolo val segment data=coco-seg.yaml batch=1 device=0|cpu
Classification (ImageNet)

See Classification Docs for usage examples with these models trained on ImageNet, which include 1000 pretrained classes.

Model size
(pixels)
acc
top1
acc
top5
Speed
CPU ONNX
(ms)
Speed
T4 TensorRT10
(ms)
params
(M)
FLOPs
(B) at 224
YOLO26n-cls 224 71.4 90.1 5.0 ± 0.3 1.1 ± 0.0 2.8 0.5
YOLO26s-cls 224 76.0 92.9 7.9 ± 0.2 1.3 ± 0.0 6.7 1.6
YOLO26m-cls 224 78.1 94.2 17.2 ± 0.4 2.0 ± 0.0 11.6 4.9
YOLO26l-cls 224 79.0 94.6 23.2 ± 0.3 2.8 ± 0.0 14.1 6.2
YOLO26x-cls 224 79.9 95.0 41.4 ± 0.9 3.8 ± 0.0 29.6 13.6
  • acc values are model accuracies on the ImageNet dataset validation set.
    Reproduce by yolo val classify data=path/to/ImageNet device=0
  • Speed averaged over ImageNet val images using an Amazon EC2 P4d instance.
    Reproduce by yolo val classify data=path/to/ImageNet batch=1 device=0|cpu
Pose (COCO)

See Pose Docs for usage examples with these models trained on COCO-Pose, which include 1 pre-trained class, person.

Model size
(pixels)
mAPpose
50-95(e2e)
mAPpose
50(e2e)
Speed
CPU ONNX
(ms)
Speed
T4 TensorRT10
(ms)
params
(M)
FLOPs
(B)
YOLO26n-pose 640 57.2 83.3 40.3 ± 0.5 1.8 ± 0.0 2.9 7.5
YOLO26s-pose 640 63.0 86.6 85.3 ± 0.9 2.7 ± 0.0 10.4 23.9
YOLO26m-pose 640 68.8 89.6 218.0 ± 1.5 5.0 ± 0.1 21.5 73.1
YOLO26l-pose 640 70.4 90.5 275.4 ± 2.4 6.5 ± 0.1 25.9 91.3
YOLO26x-pose 640 71.6 91.6 565.4 ± 3.0 12.2 ± 0.2 57.6 201.7
  • mAPval values are for single-model single-scale on COCO Keypoints val2017 dataset.
    Reproduce by yolo val pose data=coco-pose.yaml device=0
  • Speed averaged over COCO val images using an Amazon EC2 P4d instance.
    Reproduce by yolo val pose data=coco-pose.yaml batch=1 device=0|cpu
OBB (DOTAv1)

See OBB Docs for usage examples with these models trained on DOTAv1, which include 15 pre-trained classes.

Model size
(pixels)
mAPtest
50-95(e2e)
mAPtest
50(e2e)
Speed
CPU ONNX
(ms)
Speed
T4 TensorRT10
(ms)
params
(M)
FLOPs
(B)
YOLO26n-obb 1024 52.4 78.9 97.7 ± 0.9 2.8 ± 0.0 2.5 14.0
YOLO26s-obb 1024 54.8 80.9 218.0 ± 1.4 4.9 ± 0.1 9.8 55.1
YOLO26m-obb 1024 55.3 81.0 579.2 ± 3.8 10.2 ± 0.3 21.2 183.3
YOLO26l-obb 1024 56.2 81.6 735.6 ± 3.1 13.0 ± 0.2 25.6 230.0
YOLO26x-obb 1024 56.7 81.7 1485.7 ± 11.5 30.5 ± 0.9 57.6 516.5
  • mAPtest values are for single-model multiscale on DOTAv1 dataset.
    Reproduce by yolo val obb data=DOTAv1.yaml device=0 split=test and submit merged results to DOTA evaluation.
  • Speed averaged over DOTAv1 val images using an Amazon EC2 P4d instance.
    Reproduce by yolo val obb data=DOTAv1.yaml batch=1 device=0|cpu

Integrations

Our key integrations with leading AI platforms extend the functionality of Ultralytics' offerings, enhancing tasks like dataset labeling, training, visualization, and model management. Discover how Ultralytics, in collaboration with partners like Weights & Biases, Comet ML, Roboflow, and Intel OpenVINO, can optimize your AI workflow. Explore more at Ultralytics Integrations.

Ultralytics active learning integrations

Ultralytics Platform logo
Ultralytics Platform 🌟
Weights & Biases logo
Weights & Biases
Comet ML logo
Comet
Neural Magic logo
Neural Magic
Streamline Ultralytics YOLO workflows: Label, train, and deploy effortlessly with Ultralytics Platform. Try now! Track experiments, hyperparameters, and results with Weights & Biases. Free forever, Comet ML lets you save Ultralytics YOLO models, resume training, and interactively visualize predictions. Run Ultralytics YOLO inference up to 6x faster with Neural Magic DeepSparse.

Contribute

We love your input! Ultralytics YOLO would not be possible without help from our community. Please see our Contributing Guide to get started, and fill out our Survey to send us feedback on your experience. Thank you 🙏 to all our contributors!

Ultralytics open-source contributors

License

Ultralytics offers two licensing options to accommodate diverse use cases:

  • AGPL-3.0 License: This OSI-approved open-source license is ideal for students and enthusiasts, promoting open collaboration and knowledge sharing. See the LICENSE file for more details.
  • Enterprise License: Designed for commercial use, this license permits seamless integration of Ultralytics software and AI models into commercial goods and services, bypassing the open-source requirements of AGPL-3.0. If your scenario involves embedding our solutions into a commercial offering, reach out through Ultralytics Licensing.

Contact

For Ultralytics bug reports and feature requests please visit GitHub Issues. Become a member of the Ultralytics Discord, Reddit, or Forums for asking questions, sharing projects, learning discussions, or for help with all things Ultralytics!


Ultralytics GitHub Ultralytics LinkedIn Ultralytics Twitter Ultralytics YouTube Ultralytics TikTok Ultralytics BiliBili Ultralytics Discord
Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Evaluation results