How to Transform and Test the Inference Model of Forlinx RK3588 Development Board?

RKNN (Rockchip Neural Network) is a deep learning inference framework for embedded devices. It provides an end-to-end solution for transforming trained deep-learning models into executable files that run on embedded devices.The RKNN framework can efficiently run deep learning models on embedded devices, which is useful for application scenarios that require real-time inference on resource-constrained devices.

For example, RKNN can be used in smart cameras, robots, drones and other embedded devices to achieve artificial intelligence functions such as object detection, face recognition, and image classification, etc.

RK3568 Rockchip Neural Network

Let’s take an example to make the introduction: we use the RKNN-Toolkit2 tool to transform the yolov5s.onnx model in the rknpu2 project to a yolov5s.rknn model.

Development tool: Forlinx OK3588-C development board

Development environment:Ubuntu20.04

01 Download RKNN-Toolkit2

Download RKNN-Toolkit2

02 Installation Lib

requirements_cp36-1.3.0.txt is in the rknn-toolkit2/doc directory:

Installation Lib

03 Connect Development Environment to OK3588-C Development Board.

Install adb in the development environment.

Install adb in the development environment

Connect it to the TypeC0 port of the board by a USB-typeC cable. Then it’s recognized in the virtual machine by the PC.

Check connection status in the development environment.

development environment

If the connection is successful, the device ID will be sent back, as follows:

device ID

04 Download NPU Project

Download NPU Project

05 Send rknn_server and rknn library to the Development Board.

rknn_server and rknn library

Run rknn_server service on OK3588-C development board.

rknn_server service

Check the running status of rknn_server in the development environment.

running status of rknn_server

A return PID indicates successful running.

06 Model Transformation

Enter the development environment→rknn_toolkit2→examples directory→select a model. In this example, we choose to transform the onnx model to an RKNN model.



Add target_platform='rk3588' to rknn.config

Add target='rk3588' to rknn.init_runtime
Run after modification.


A successful run is as follows:

successful run

successful run

As well the yolov5s.rknn model is generated in the directory.

yolov5s.rknn model

07 Compile and Test the Source Code

Go to the dirctory: rknpu2/examples/rknn_yolov5_demo and set the environment variables:


Execute the compile script to compile:

compile script

Then rknn_yolov5_demo is generated in the directory: rknpu2/examples/rknn_yolov5_demo/install/rknn_yolov5_demo_Linux


08 Test

Copy the yolov5s.rknn model generated above and the rknn_yolov5_demo_Linux in the install directory to the development board.

yolov5s.rknn model

Go to the directory: rknn_yolov5_demo_Linux and add the environment variables for the linked libraries (lib directory in the rknn_yolov5_demo_Linux directory)

linked libraries

Use rknn model to conduct object recognition commands as follows:

rknn model

Execution results:

Execution results

Copy the generated out.jpg to the local computer to review the recognition results:

recognition results

That’s all the procedure of transforming and testing the RK3588 inference model based on Forlinx OK3588-C development board.