0

$USD

$USD
TUTORIALS

Running YOLOv10 on Unihiker: A Step-by-Step Guide for Efficient Object Detection

DFRobot Aug 14 2024 343

For developers, educators, and hobbyists interested in deploying advanced object detection algorithms on compact hardware platforms, this guide offers a comprehensive overview. If you're looking to explore the integration of AI with embedded systems, especially using the YOLOv10 algorithm on the UNIHIKER - IoT Python Single Board Computer, you'll find it particularly useful. You'll learn how to set up the environment, optimize the model for performance, and implement practical applications. This makes it a valuable resource for enhancing your skills in both AI and embedded system development.

 

1. Introduction

1.1 Introduction to YOLOv10

YOLO (You Only Look Once) series is currently one of the mainstream end-side object detection algorithms, first proposed by Joseph Redman and others. Over time, multiple versions have been released, each version “seemingly” improving in performance and speed.

This article introduces YOLOv10 running on the Unihiker board. YOLOv10, proposed by the research team from Tsinghua University, follows the design principles of the YOLO series and is dedicated to creating a real-time end-to-end high-performance object detector. YOLOv10 addresses the shortcomings of the YOLO series in post-processing and model architecture. By eliminating non-maximum suppression (NMS) operations and optimizing the model architecture, YOLOv10 significantly reduces computational overhead while achieving state-of-the-art performance. Extensive experiments on standard object detection benchmarks show that YOLOv10 significantly outperforms previous state-of-the-art models in terms of computation-accuracy trade-offs across various model scales. As shown in the figure below, YOLOv10-S / X is 1.8 times / 1.3 times faster than RT-DETR R18 / R101 with similar performance. Compared to YOLOv9-C, YOLOv10-B achieves a 46% reduction in latency with the same performance. Additionally, YOLOv10 demonstrates extremely high parameter utilization efficiency. YOLOv10-L / X has 1.8 times and 2.3 times fewer parameters, respectively, outperforming YOLOv8-L / X by 0.3 AP and 0.5 AP. YOLOv10-M, with 23% and 31% fewer parameters, achieves similar AP to YOLOv9-M / YOLO-MS.

Comparison of yolov10 performance parameters

Comparison of yolov10 performance parameters

 

1.2 Introduction to Unihiker Board Hardware

The UNIHIKER - IoT Python Single Board Computer is a new generation of domestically produced open-source hardware designed specifically for Python learning and usage. It adopts a single-board computer architecture and integrates an LCD color screen, WiFi Bluetooth, various common sensors, and a rich set of expansion interfaces. Additionally, it comes with a Linux operating system and Python environment pre-installed with common Python libraries, allowing students and teachers to start Python teaching in just two steps.

Unihiker Board

 

The Unihiker Board is based on the RK3308 Arm 64-bit quad-core processor with a clock speed of 1.2GHz, equipped with 512MB DDR3 memory and 16GB eMMC hard drive, running the Debian 10 operating system. It also supports 2.4G Wi-Fi and Bluetooth 4.0, using the RTL8723DS chip. The Unihiker Board also integrates a GD32VF103C8T6 RISC-V coprocessor with a clock speed of 108MHz, 64KB Flash, and 32KB SRAM.

The Unihiker Board has various onboard components, including a Home button, A/B buttons, a 2.8-inch touch-enabled color screen with a resolution of 240x320, a capacitive silicon microphone, a PT0603 photo-sensitive triode light sensor, a passive buzzer, and a blue LED. It also has an ICM20689 six-axis sensor that includes a three-axis accelerometer and a three-axis gyroscope.

In terms of interfaces, the Unihiker Board offers multiple connection options. It has a USB Type-C port for connecting the CPU to a PC for programming or powering the main board. There is also a USB TYPE-A port for connecting external USB devices. Additionally, the board features a microSD card interface for expanding storage space, a 3Pin I/O supporting three 10-bit PWM channels and two 12-bit ADC channels, a separate 4Pin I2C interface, and 19 independent I/O gold fingers compatible with micro:bit, supporting various communication protocols and functions.

Unihiker Board offers multiple connection options

 

1.3 Object Detection with YOLOv10 on the Unihiker Board

In this article, we will use the Unihiker Board developed by DFRobot to run YOLOv10 and attempt to accelerate it by converting to ONNX format.

Deploying YOLOv10 on the Unihiker Board has significant practical and educational implications:

1. Portability and deployment flexibility: The compact size of the Unihiker Board makes it suitable for embedding into space-constrained devices, enabling portable object detection deployments. Compared to large computers, the Unihiker Board is more suited for on-site and mobile scenarios.

2. Cost-effectiveness: The Unihiker Board is relatively low-cost, suitable for budget-conscious projects and educational purposes. By running YOLOv10 on the Unihiker Board, low-cost object detection application development and experimentation can be conducted.

3. Learning and experimental platform: The Unihiker Board provides a wealth of interfaces and onboard components, making it suitable as a learning and experimental platform. By running YOLOv10 on the Unihiker Board, students and developers can gain a deep understanding of the integration of embedded systems and artificial intelligence, learning about optimization and acceleration algorithms in resource-constrained environments.

4. Technical challenges and innovation: Running YOLOv10 on the resource-limited Unihiker Board involves overcoming challenges related to computational performance and memory limitations. This provides developers with an opportunity to explore and innovate, trying various optimization techniques such as model compression, quantization, and hardware acceleration.

 

2. Preparation: Setting up the Environment for Running YOLOv10

To successfully run YOLOv10 on the Unihiker Board, we will use the library provided by Ultralytics for deployment. First, we need to ensure that the Python environment on the Unihiker Board meets the requirements for YOLOv10, specifically upgrading to Python version 3.8 or higher. For this purpose, we recommend using MiniConda for version management, allowing easy switching and management of different Python environments.

The steps are as follows:

1. Install MiniConda: First, download and install MiniConda on the Unihiker Board. MiniConda is a lightweight Python distribution specifically designed to simplify the management of Python environments and package installations.

2. Create a new environment: Use MiniConda to create a new virtual environment with Python 3.8 or higher. This ensures that our deployment of YOLOv10 is not affected by the default system Python environment, avoiding compatibility issues.

3. Activate the environment: Activate the newly created virtual environment to make it the current working environment.

4. Install the Ultralytics library: In the activated virtual environment, install the YOLO library provided by Ultralytics using the pip command. This will download and install all necessary dependencies and components, enabling us to smoothly run YOLOv10.

By following these steps, we can successfully deploy YOLOv10 on the Unihiker Board, fully utilizing its powerful object detection capabilities. This method not only ensures efficient operation of YOLOv10 but also facilitates environment management and version control, providing a stable foundation for subsequent development and experimentation. Additionally, by using MiniConda for version management, we can more flexibly address different projects' Python environment requirements, improving development efficiency.

 

Below are the detailed steps:

Step 1 Check the current Python version:

In the terminal, enter:

python --version

The terminal displays:

Python 3.7.3

Ultralytics does not support lower versions of Python; thus, it is necessary to upgrade Python. We choose to use MiniConda for version management and upgrading.

Note: Do not use Anaconda, as it may cause errors when running on the Unihiker Board.

 

Step 2 Download MiniConda:

In the terminal, enter:

wget https://github.com/conda-forge/miniforge/releases/latest/download/Miniforge3-Linux-aarch64.sh

After the download is complete, the terminal displays:

Saved “Miniforge3-Linux-aarch64.sh” [74300552/74300552])

 

Step 3 Install MiniConda:

In the terminal, enter:

sudo bash Miniforge3-Linux-aarch64.sh

Follow the prompts to press ENTER or yes keys when encountered. The terminal finally displays:

Added mamba to /root/.bashrc

==> For changes to take effect, close and re-open your current shell. <==

Thank you for installing Miniforge3!

Install MiniConda

In the terminal, enter:

source ~/.bashrc

After installation, enter in the terminal:

conda

The terminal displays:

Install MiniConda

 

Step 4 Activate Conda:

In the terminal, enter:

conda activate

The terminal display changes from:

root@unihiker:

to:

(base)root@unihiker:

indicating successful activation of Conda.

 

Step 5 Create a YOLO environment in Conda:

Name the environment "yolo" and choose Python version 3.11. In the terminal, enter:

conda create -n yolo python==3.11

During the process, the display shows:

Create a YOLO environment in Conda

Enter 'y'

After the environment is created, the terminal displays:

Create a YOLO environment in Conda

 

Step 6 Activate the YOLO environment:

In the terminal, enter:

conda activate yolo

The terminal display changes from:

(base)root@unihiker:

to:

(yolo)root@unihiker:

indicating successful activation of the YOLO environment.

 

Step 7 Install Ultralytics:

In the terminal, enter:

pip install ultralytics

After completion, the terminal displays:

Install Ultralytics

 

Step 8 Install the Pillow library:

In the terminal, enter:

pip install pillow

If already installed, the terminal displays:

Requirement already satisfied: pillow in /root/miniforge3/envs/yolo/lib/python3.11/site-packages (10.3.0)

 

Step 9 Install OpenCV:

In the terminal, enter:

pip install opencv-python

If already installed, the terminal displays:

Requirement already satisfied: opencv-python in /root/miniforge3/envs/yolo/lib/python3.11/site-packages (4.9.0.80)

Requirement already satisfied: numpy>=1.21.2 in /root/miniforge3/envs/yolo/lib/python3.11/site-packages (from opencv-python) (1.26.4)

 

Step 10 Install the Hugging Face library:

In the terminal, enter:

pip install huggingface

Install the Hugging Face library

 

Step 11 Install the Hugging Face Hub library:

In the terminal, enter:

pip install huggingface_hub

Install the Hugging Face Hub library

 

3. Quick Start: Running Native YOLOv10

Step 1 Download the project from GitHub:

YOLOv10 project URL: https://github.com/THU-MIG/yolov10

In the terminal, enter:

git clone https://github.com/THU-MIG/yolov10.git

Download the YOLOv10 project. Then enter the directory by typing in the terminal:

cd yolov10

Step 2 Prepare the required files:

Weight files:

[yolov10n.pt]

 

Step 3 Create a Python file:

quick_start.py

Sample code:

 

Python
from ultralytics import YOLO

# Load a pretrained YOLO model (recommended for training)
model = YOLO("yolov10n.pt")

# Perform object detection on an image using the model
results = model("https://ultralytics.com/images/bus.jpg")

# Save results to disk
results[0].save(filename=f"result_bus.jpg")

 

Step 4 Confirm YOLO environment is activated:

In the terminal, enter:

conda activate yolo

Confirm the terminal display:

Confirm YOLO environment is activated

 

Step 5 Run the Python script you wrote:

In the terminal, enter:

python quick_start.py

The terminal display shows:

Run the Python script you wrote

Using the native YOLOv10n model for single image inference takes about 7 seconds.

The image "bus.jpg" was automatically downloaded, which we prepared for YOLO inference, as shown below:

Run the Python script you wrote

Finally, the model inference results are stored as "result_bus.jpg":

Run the Python script you wrote

 

4. Optimization: Converting to ONNX Format

When running the native YOLOv10n, the speed is quite slow, so we need to convert it to ONNX format to speed up its operation.

This section describes how to convert yolov10n.pt to ONNX format to accelerate its operation. Here are the detailed steps:

 

Step 1 Move to the YOLO directory created in the previous section:

In the terminal, enter:

cd yolov10

 

Step 2 Create a Python file named export_onnx.py with the following code:

Python
from ultralytics import YOLO

# Load a model
model = YOLO("yolov10n.pt")  # load an official model

# Export the model
model.export(format="onnx")

Step 3 Confirm the YOLO environment is activated:

In the terminal, enter:

conda activate yolo

Confirm the terminal display:

Confirm the YOLO environment is activated

 

Step 4 Run the Python script you wrote:

In the terminal, enter:

python export_onnx.py

The terminal displays:

Run the Python script you wrote

You can see that the converted file "yolov10n.onnx" was automatically saved.

 

Step 5 Write a file to run the ONNX model:

Create a file named predict_onnx.py and write the following code:

Python
from ultralytics import YOLO

# Load the exported NCNN model
onnx_model = YOLO("yolov10n.onnx", task = 'detect')

# Run inference
results = onnx_model("https://ultralytics.com/images/bus.jpg")

# Save results to disk
results[0].save(filename=f"result_bus_onnx.jpg")

Step 6 Run the inference code:

In the terminal, enter:

python predict_onnx.py

The terminal displays:

Run the inference code

The directory generates the prediction result:

Run the inference code

You can see that using the ONNX model for object detection takes about 6.5 seconds, which is 0.5 seconds faster than the native model.

 

5. Further Optimization: Reducing the Input Image Size

It is evident that the current inference speed is still slow.

If you want to improve inference speed, you can reduce the input image size. When exporting the ONNX model, we can set the imgsz parameter to specify the input image size. If the input image size is uncertain, we can also set the dynamic parameter to True, allowing the exported ONNX to accept any size of image input for inference. Here are the specific steps:

Step 1 Move to the YOLO directory created in the previous section:

In the terminal, enter:

cd yolov10

 

Step 2 Create a Python file named export_onnx.py with the following code:

In this step, we set the dynamic parameter to True.

Python
from ultralytics import YOLO

# Load a model
model = YOLO("yolov10n.pt")  # load an official model

# Export the model
model.export(format="onnx", dynamic = True)

Step 3 Confirm the YOLO environment is activated:

In the terminal, enter:

conda activate yolo

Confirm the terminal display:

Confirm the YOLO environment is activated

 

Step 4 Run the Python script you wrote:

In the terminal, enter:

python export_onnx.py

The terminal displays:

Run the Python script you wrote

You can see that the file "yolov10n.onnx" was automatically saved.

 

Step 5 Write a file to run the ONNX model:

Create a file named predict_onnx.py and write the following code:

Python
from ultralytics import YOLO
import cv2

# Load the exported NCNN model
onnx_model = YOLO("yolov10n.onnx", task = 'detect')

image = cv2.imread('bus.jpg')
print(image.shape)

# Run inference
print('original')
results = onnx_model("bus.jpg")
results[0].save(filename='bus_640.jpg')
print(256)
results = onnx_model("bus.jpg", imgsz = 256)
results[0].save(filename='bus_256.jpg')
print(128)
results = onnx_model("bus.jpg", imgsz = 128)
results[0].save(filename='bus_128.jpg')
print(64)
results = onnx_model("bus.jpg", imgsz = 64)
results[0].save(filename='bus_64.jpg')

In this code, we test the running conditions of the original size image, 256 size image, 128 size image, and 64 size image.

 

Step 6 Run the inference code:

In the terminal, enter:

python predict_onnx.py

The terminal displays:

Run the inference code

The original image size is 1080*810, and the maximum prediction size of the native YOLOv10n is 640, which takes about 6 seconds. The results are as follows:

Run the inference code

When the input size is 448, the time taken is about seconds. The results are as follows:

Run the inference code

When the input size is 320, the time taken is about 1.2 seconds. The results are as follows:

Run the inference code

When the input size is 256, the time taken is about 0.8 seconds.

Run the inference code

When the input size is 128, the time taken is about 0.4 seconds.

Run the inference code

In summary:
 Note: Performance is based on the bus.jpg image. If you have your own dataset, further testing may be needed.

In this test, the accuracy judgment standard is:

AccuracyVery GoodGoodAveragePoorVery Poor
Difference in the number of objects recognized compared to Ground Truth0123>3
SizeTime TakenAccuracy
6406sVery Good
4482.5sVery Good
3201.2sGood
2560.7sGood
1280.28sAverage
64UnsupportedUnsupported

 

6. Comparison with YOLOv8n

Using official code, the comparison of image size and running time is as follows:

Note: Accuracy is tested with bus.jpg. If you have your own dataset, further testing may be needed.

Sizeyolov8nyolov10n
Time TakenAccuracyTime TakenAccuracy
64022sVery Good6sVery Good
4483.5sVery Good2.5sVery Good
3202.2sVery Good1.2sGood
2560.8sVery Good0.7sGood
1280.4sGood0.28sAverage
640.1sPoorUnsupportedUnsupported

7. Conclusion

The Unihiker Board can run YOLOv10n

Using the Unihiker Board, YOLOv10n can be run with simple code, and the speed is slightly faster than YOLOv8n.

 

This article introduces how to complete the deployment of YOLOv10 on the Unihiker Board

Although YOLOv10 is developed based on Ultralytics, simply installing the Ultralytics library does not allow direct running of YOLOv10; it is still necessary to clone the official repository.

 

Recommended resolution configuration

If using the Unihiker Board for image object detection, consider using an input resolution of 448, where the processing time for one image is about 2.5 seconds, while the accuracy is very good. If using the Unihiker Board for video object detection, given the computational power of the Unihiker Board, it is recommended to use an input resolution of 128, where the processing time for one image is about 0.28 seconds, but the performance is average at this time. It is worth considering using YOLOv8n's input resolution of 128, where the accuracy performance is better. Again, accuracy performance is based on bus.jpg testing. If you have your own dataset, further testing may be needed.

 

Continuous optimization, stay tuned

We will continue to optimize the running of YOLOv10n on the Unihiker Board, stay tuned.

REVIEW