Back to Ultralytics

Ultralytics YOLO26 Modes

docs/en/modes/index.md

8.4.4610.2 KB
Original Source

Ultralytics YOLO26 Modes

Introduction

Ultralytics YOLO26 is not just another object detection model; it's a versatile framework designed to cover the entire lifecycle of machine learning models—from data ingestion and model training to validation, deployment, and real-world tracking. Each mode serves a specific purpose and is engineered to offer you the flexibility and efficiency required for different tasks and use cases.

<p align="center"> <iframe loading="lazy" width="720" height="405" src="https://www.youtube.com/embed/j8uQc0qB91s?si=dhnGKgqvs7nPgeaM" title="YouTube video player" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" allowfullscreen> </iframe>

<strong>Watch:</strong> Ultralytics Modes Tutorial: Train, Validate, Predict, Export & Benchmark.

</p>

Modes at a Glance

Understanding the different modes that Ultralytics YOLO26 supports is critical to getting the most out of your models:

  • Train mode: Fine-tune your model on custom or preloaded datasets.
  • Val mode: A post-training checkpoint to validate model performance.
  • Predict mode: Unleash the predictive power of your model on real-world data.
  • Export mode: Make your model deployment-ready in various formats.
  • Track mode: Extend your object detection model into real-time tracking applications.
  • Benchmark mode: Analyze the speed and accuracy of your model in diverse deployment environments.

This comprehensive guide aims to give you an overview and practical insights into each mode, helping you harness the full potential of YOLO26.

Train

Train mode is used for training a YOLO26 model on a custom dataset. In this mode, the model is trained using the specified dataset and hyperparameters. The training process involves optimizing the model's parameters so that it can accurately predict the classes and locations of objects in an image. Training is essential for creating models that can recognize specific objects relevant to your application.

Train Examples{ .md-button }

Val

Val mode is used for validating a YOLO26 model after it has been trained. In this mode, the model is evaluated on a validation set to measure its accuracy and generalization performance. Validation helps identify potential issues like overfitting and provides metrics such as mean Average Precision (mAP) to quantify model performance. This mode is crucial for tuning hyperparameters and improving overall model effectiveness.

Val Examples{ .md-button }

Predict

Predict mode is used for making predictions using a trained YOLO26 model on new images or videos. In this mode, the model is loaded from a checkpoint file, and the user can provide images or videos to perform inference. The model identifies and localizes objects in the input media, making it ready for real-world applications. Predict mode is the gateway to applying your trained model to solve practical problems.

Predict Examples{ .md-button }

Export

Export mode is used for converting a YOLO26 model to formats suitable for deployment across different platforms and devices. This mode transforms your PyTorch model into optimized formats like ONNX, TensorRT, or CoreML, enabling deployment in production environments. Exporting is essential for integrating your model with various software applications or hardware devices, often resulting in significant performance improvements.

Export Examples{ .md-button }

Track

Track mode extends YOLO26's object detection capabilities to track objects across video frames or live streams. This mode is particularly valuable for applications requiring persistent object identification, such as surveillance systems or self-driving cars. Track mode implements sophisticated algorithms like ByteTrack to maintain object identity across frames, even when objects temporarily disappear from view.

Track Examples{ .md-button }

Benchmark

Benchmark mode profiles the speed and accuracy of various export formats for YOLO26. This mode provides comprehensive metrics on model size, accuracy (mAP50-95 for detection tasks or accuracy_top5 for classification), and inference time across different formats like ONNX, OpenVINO, and TensorRT. Benchmarking helps you select the optimal export format based on your specific requirements for speed and accuracy in your deployment environment.

Benchmark Examples{ .md-button }

FAQ

How do I train a custom object detection model with Ultralytics YOLO26?

Training a custom object detection model with Ultralytics YOLO26 involves using the train mode. You need a dataset formatted in YOLO format, containing images and corresponding annotation files. Use the following command to start the training process:

!!! example

=== "Python"

    ```python
    from ultralytics import YOLO

    # Load a pretrained YOLO model (you can choose n, s, m, l, or x versions)
    model = YOLO("yolo26n.pt")

    # Start training on your custom dataset
    model.train(data="path/to/dataset.yaml", epochs=100, imgsz=640)
    ```

=== "CLI"

    ```bash
    # Train a YOLO model from the command line
    yolo detect train data=path/to/dataset.yaml model=yolo26n.pt epochs=100 imgsz=640
    ```

For more detailed instructions, you can refer to the Ultralytics Train Guide.

What metrics does Ultralytics YOLO26 use to validate the model's performance?

Ultralytics YOLO26 uses various metrics during the validation process to assess model performance. These include:

  • mAP (mean Average Precision): This evaluates the accuracy of object detection.
  • IOU (Intersection over Union): Measures the overlap between predicted and ground truth bounding boxes.
  • Precision and Recall: Precision measures the ratio of true positive detections to the total detected positives, while recall measures the ratio of true positive detections to the total actual positives.

You can run the following command to start the validation:

!!! example

=== "Python"

    ```python
    from ultralytics import YOLO

    # Load a pretrained or custom YOLO model
    model = YOLO("yolo26n.pt")

    # Run validation on your dataset
    model.val(data="path/to/validation.yaml")
    ```

=== "CLI"

    ```bash
    # Validate a YOLO model from the command line
    yolo val model=yolo26n.pt data=path/to/validation.yaml
    ```

Refer to the Validation Guide for further details.

How can I export my YOLO26 model for deployment?

Ultralytics YOLO26 offers export functionality to convert your trained model into various deployment formats such as ONNX, TensorRT, CoreML, and more. Use the following example to export your model:

!!! example

=== "Python"

    ```python
    from ultralytics import YOLO

    # Load your trained YOLO model
    model = YOLO("yolo26n.pt")

    # Export the model to ONNX format (you can specify other formats as needed)
    model.export(format="onnx")
    ```

=== "CLI"

    ```bash
    # Export a YOLO model to ONNX format from the command line
    yolo export model=yolo26n.pt format=onnx
    ```

Detailed steps for each export format can be found in the Export Guide.

What is the purpose of the benchmark mode in Ultralytics YOLO26?

Benchmark mode in Ultralytics YOLO26 is used to analyze the speed and accuracy of various export formats such as ONNX, TensorRT, and OpenVINO. It provides metrics like model size, mAP50-95 for object detection, and inference time across different hardware setups, helping you choose the most suitable format for your deployment needs.

!!! example

=== "Python"

    ```python
    from ultralytics.utils.benchmarks import benchmark

    # Run benchmark on GPU (device 0)
    # You can adjust parameters like model, dataset, image size, and precision as needed
    benchmark(model="yolo26n.pt", data="coco8.yaml", imgsz=640, half=False, device=0)
    ```

=== "CLI"

    ```bash
    # Benchmark a YOLO model from the command line
    # Adjust parameters as needed for your specific use case
    yolo benchmark model=yolo26n.pt data='coco8.yaml' imgsz=640 half=False device=0
    ```

For more details, refer to the Benchmark Guide.

How can I perform real-time object tracking using Ultralytics YOLO26?

Real-time object tracking can be achieved using the track mode in Ultralytics YOLO26. This mode extends object detection capabilities to track objects across video frames or live feeds. Use the following example to enable tracking:

!!! example

=== "Python"

    ```python
    from ultralytics import YOLO

    # Load a pretrained YOLO model
    model = YOLO("yolo26n.pt")

    # Start tracking objects in a video
    # You can also use live video streams or webcam input
    model.track(source="path/to/video.mp4")
    ```

=== "CLI"

    ```bash
    # Perform object tracking on a video from the command line
    # You can specify different sources like webcam (0) or RTSP streams
    yolo track model=yolo26n.pt source=path/to/video.mp4
    ```

For in-depth instructions, visit the Track Guide.