- Last edited 3 months ago ago
How to compile model and run inference on Coral Edge TPU using STM32MP1
- 1 Article purpose
- 2 Prerequisites
- 3 Fetching a Coral Edge TPU compiled model
- 4 Running the inference
- 5 References
1 Article purpose
This article aims at describing how to run an inference on Coral Edge TPU using STM32MP1 microprocessor devices.
2.1 Installing the Edge TPU compiler
To perform an inference on the Coral Edge TPU hardware, we need to convert our TensorFlow Lite model into an Edge TPU model. This can be achieved by using the Edge TPU compiler. To do this, install the Edge TPU compiler on your host computer.
Install the Coral Edge TPU compiler by running the following commands on your host computer:
curl https://packages.cloud.google.com/apt/doc/apt-key.gpg | sudo apt-key add - echo "deb https://packages.cloud.google.com/apt coral-edgetpu-stable main" | sudo tee /etc/apt/sources.list.d/coral-edgetpu.list sudo apt-get update sudo apt-get install edgetpu-compiler
Enter the following command to check that the compiler has been installed:
edgetpu_compiler -v Edge TPU Compiler version x.y.z
3 Fetching a Coral Edge TPU compiled model
3.1 Coral compiled models
3.2 Compiling your own model
The Edge TPU compiler role is to convert one or several TensorFlow Lite models into Edge TPU compatible models. It takes as an argument your .tflite model and returns a .tflite Edge TPU compatible model. You can pass multiple models as arguments (each separated with a space). They are then co-compiled and ready to share the Edge TPU 8 Mbytes of RAM for parameter data caching.
Be aware that not all the operations supported on TensorFlow Lite are supported on Edge TPU. While building your own model architecture, check the operations and the layers supported by the Edge TPU compiler on Coral AI supported operations .
If your model architecture uses unsupported operations and does not meet all the requirements, then only the first portion of the model will be executed on the Edge TPU compiler. Starting from the first node in your model graph where an unsupported operation occurs, all operations are run on the CPU of the target board even if an Edge TPU supported operation occurs. The Edge TPU compiler cannot partition the model more than once.
To compile your .tflite model, execute the following command:
edgetpu_compiler -m 13 your_model_1.tflite
3.3 Example: compiling an object detection model
The object detection model used is the ssd_mobilenet_v1_coco_quant.tflite downloaded from the Coral  website and compiled for the Coral Edge TPU using the steps detailed below:
wget http://storage.googleapis.com/download.tensorflow.org/models/tflite/coco_ssd_mobilenet_v1_1.0_quant_2018_06_29.zip unzip ./coco_ssd_mobilenet_v1_1.0_quant_2018_06_29.zip Archive: ./coco_ssd_mobilenet_v1_1.0_quant_2018_06_29.zip inflating: detect.tflite inflating: labelmap.txt
Now that the model has been downloaded and extracted successfully, it is time to compile it using the Edge TPU compiler on your host PC:
edgetpu_compiler -m 13 detect.tflite
3.4 Sending your model to the target
In your board workspace directory, create two main directories to organize the workflow:
cd /usr/local && mkdir -p workspace cd /usr/local/workspace && mkdir -p models
Then transfer your compiled model from the host computer to the models directory in the board workspace:
scp path/to/your/compiled_model_edgetpu.tflite root@<board_ip_address>:/usr/local/workspace/models/
Now that your workspace is ready with the compiled model file, it is time to see how to run an inference using the C++ benchmark application.
4 Running the inference
4.1 Installing the benchmark application
Configure the AI OpenSTLinux package, then install X-LINUX-AI components for this application:
apt-get install tflite-edgetpu-benchmark
4.2 Execute the benchmark on the model
Now that our compiled model is loaded on the board, it is time to run an inference using the benchmark example. This example aims at measuring your model average inference time on a predefined number of 25 loops. To do this, execute the following command:
cd /usr/local/demo-ai/benchmark/tflite-edgetpu/ ./tflite_edgetpu_benchmark -m /usr/local/workspace/models/your_compiled_model.tflite -l 25
The first inference may take longer since the model is being loaded on the Coral Edge TPU RAM. This time is not taken into account in the average inference time.
4.3 Customizing your application
You can adapt your application to your development constraints and requirements.