HTVM: Efficient Neural Network Deployment On Heterogeneous TinyML Platforms

June 11, 2024 ยท Entered Twilight ยท ๐Ÿ› Design Automation Conference

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: .asf.yaml, .clang-format, .gitattributes, .github, .gitignore, .gitmodules, .pre-commit-config.yaml, 3rdparty, CMakeLists.txt, CONTRIBUTORS.md, Jenkinsfile, KEYS, LICENSE, Makefile, NEWS.md, NOTICE, README.md, apps, ci, cmake, cmd.txt, conda, configs, conftest.py, diana, docker, docs, gallery, golang, include, jvm, licenses, mypy.ini, pyproject.toml, python, rust, src, tests, version.py, vta, web

Authors Josse Van Delm, Maarten Vandersteegen, Alessio Burrello, Giuseppe Maria Sarda, Francesco Conti, Daniele Jahier Pagliari, Luca Benini, Marian Verhelst arXiv ID 2406.07453 Category cs.PL: Programming Languages Cross-listed cs.DC Citations 12 Venue Design Automation Conference Repository https://github.com/KULeuven-MICAS/htvm โญ 16 Last Checked 2 months ago
Abstract
Optimal deployment of deep neural networks (DNNs) on state-of-the-art Systems-on-Chips (SoCs) is crucial for tiny machine learning (TinyML) at the edge. The complexity of these SoCs makes deployment non-trivial, as they typically contain multiple heterogeneous compute cores with limited, programmer-managed memory to optimize latency and energy efficiency. We propose HTVM - a compiler that merges TVM with DORY to maximize the utilization of heterogeneous accelerators and minimize data movements. HTVM allows deploying the MLPerf(TM) Tiny suite on DIANA, an SoC with a RISC-V CPU, and digital and analog compute-in-memory AI accelerators, at 120x improved performance over plain TVM deployment.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Programming Languages