Skip to content

Extension package of Apache TVM (Machine Learning Compiler) for Renesas DRP-AI accelerators powered by Edgecortix MERA(TM) Based Apache TVM version: v0.11.1

License

Notifications You must be signed in to change notification settings

renesas-rz/rzv_drp-ai_tvm

Repository files navigation

Extension package of TVM Deep Learning Compiler for Renesas DRP-AI accelerators powered by EdgeCortix MERA™ (DRP-AI TVM)

TVM Documentation | TVM Community | TVM github |

DRP-AI TVM 1 is Machine Learning Compiler plugin for Apache TVM with AI accelerator DRP-AI provided by Renesas Electronics Corporation.

License

(C) Copyright EdgeCortix, Inc. 2022 (C) Copyright Renesas Electronics Corporation 2022 Contributors Licensed under an Apache-2.0 license.

Supported Embedded Platforms

  • Renesas RZ/V2L Evaluation Board Kit (How to get)
  • Renesas RZ/V2M Evaluation Board Kit (How to get)
  • Renesas RZ/V2MA Evaluation Board Kit (How to get)
  • Renesas RZ/V2H Evaluation Board Kit

Introduction

Overview

This compiler stack is an extension of the DRP-AI Translator to the TVM backend. CPU and DRP-AI can work together for the inference processing of the AI models. For more information on the features of DRP-AI TVM and various content (documents, demo videos, training videos, GUI demos, etc.), please refer to the DRP-AI TVM on RZ/V Series (GitHub.IO) page.

File Configuration

Directory Details
tutorials Sample compile script
apps Sample inference application on the target board
setup Setup scripts for building a TVM environment
obj Pre-build runtime binaries
docs "Model List", "Error List", documentation of TVM API wrapper for sample codes, and etc.
img Image files used in this document
tvm TVM repository from GitHub
3rdparty 3rd party tools
how-to Sample to solve specific problems, i.e., How to run validation between x86 and DRP-AI
pruning Sample scripts to prune the models with DRP-AI Extension Package

Installation

Deploy AI models on DRP-AI

Video

Following video shows brief tutorial for how to deploy ONNX model on RZ/V series board.
RZ/V DRP-AI TVM Tutorial - How to Run ONNX Model (YouTube)

Tutorial Video

Overview

To deploy the AI model to DRP-AI on the target board, you need to compile the model with DRP-AI TVM1 to generate Runtime Model Data (Compile).
SDK generated from RZ/V Linux Package and DRP-AI Support Package is required to compile the model.

After compiled the model, you need to copy the file to the target board (Deploy).
You also need to copy the C++ inference application and DRP-AI TVM1 Runtime Library to run the AI model inference.

RZ/V2L, RZ/V2M, RZ/V2MA
RZ/V2H

Following pages show the example to compile the ResNet18 model and run it on the target board.

[Option] Prune the model with DRP-AI Extension Package

Compile model with DRP-AI TVM1

Run inference on board

Sample Application

To find more AI examples, please see How-to page.

Confirmed AI Model

If you want to know which models have been tested by Renesas, please refer to Model List.

Support

Error List

If error occurred at compile/runtime operation, please refer error list.

Tips

How-to page includes following explanation.

  • profiler;
  • validation between x86 and DRP-AI;
  • etc.

Enquiries

If you have any questions, please contact Renesas Technical Support.

Footnotes

  1. DRP-AI TVM is powered by EdgeCortix MERA™ Compiler Framework. 2 3 4

About

Extension package of Apache TVM (Machine Learning Compiler) for Renesas DRP-AI accelerators powered by Edgecortix MERA(TM) Based Apache TVM version: v0.11.1

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •