Zur Kurzanzeige

dc.contributor.author
Burrello, Alessio
dc.contributor.author
Scherer, Moritz
dc.contributor.author
Zanghieri, Marcello
dc.contributor.author
Conti, Francesco
dc.contributor.author
Benini, Luca
dc.date.accessioned
2021-09-30T07:45:12Z
dc.date.available
2021-09-30T02:39:54Z
dc.date.available
2021-09-30T07:45:12Z
dc.date.issued
2021
dc.identifier.isbn
978-1-6654-3156-9
en_US
dc.identifier.isbn
978-1-6654-3157-6
en_US
dc.identifier.other
10.1109/COINS51742.2021.9524173
en_US
dc.identifier.uri
http://hdl.handle.net/20.500.11850/507694
dc.description.abstract
Transformer networks have become state-of-The-Art for many tasks such as NLP and are closing the gap on other tasks like image recognition. Similarly, Transformers and Attention methods are starting to attract attention on smaller-scale tasks, which fit the typical memory envelope of MCUs. In this work, we propose a new set of execution kernels tuned for efficient execution on MCU-class RISC-V and ARM Cortex-M cores. We focus on minimizing memory movements while maximizing data reuse in the Attention layers. With our library, we obtain 3.4×, 1.8×, and 2.1× lower latency and energy on 8-bit Attention layers, compared to previous state-of-The-Art (SoA) linear and matrix multiplication kernels in the CMSIS-NN and PULP-NN libraries on the STM32H7 (Cortex M7), STM32L4 (Cortex M4), and GAP8 (RISC-V IMC-Xpulp) platforms, respectively. As a use case for our TinyTransformer library, we also demonstrate that we can fit a 263 kB Transformer on the GAP8 platform, outperforming the previous SoA convolutional architecture on the TinyRadarNN dataset, with a latency of 9.24 ms and 0.47 mJ energy consumption and an accuracy improvement of 3.5%.
en_US
dc.language.iso
en
en_US
dc.publisher
IEEE
en_US
dc.subject
TinyML
en_US
dc.subject
Transformers
en_US
dc.subject
Deep learning
en_US
dc.subject
Internet of Things
en_US
dc.title
A Microcontroller is All You Need: Enabling Transformer Execution on Low-Power IoT Endnodes
en_US
dc.type
Conference Paper
dc.date.published
2021-09-02
ethz.book.title
2021 IEEE International Conference on Omni-Layer Intelligent Systems (COINS)
en_US
ethz.pages.start
9524173
en_US
ethz.size
6 p.
en_US
ethz.event
2021 IEEE International Conference on Omni-Layer Intelligent Systems (IEEE COINS 2021)
en_US
ethz.event.location
Online
en_US
ethz.event.date
August 23-25, 2021
en_US
ethz.identifier.scopus
ethz.publication.place
Piscataway, NJ
en_US
ethz.publication.status
published
en_US
ethz.leitzahl
ETH Zürich::00002 - ETH Zürich::00012 - Lehre und Forschung::00007 - Departemente::02140 - Dep. Inf.technologie und Elektrotechnik / Dep. of Inform.Technol. Electrical Eng.::02636 - Institut für Integrierte Systeme / Integrated Systems Laboratory::03996 - Benini, Luca / Benini, Luca
ethz.leitzahl.certified
ETH Zürich::00002 - ETH Zürich::00012 - Lehre und Forschung::00007 - Departemente::02140 - Dep. Inf.technologie und Elektrotechnik / Dep. of Inform.Technol. Electrical Eng.::02636 - Institut für Integrierte Systeme / Integrated Systems Laboratory::03996 - Benini, Luca / Benini, Luca
ethz.date.deposited
2021-09-30T02:39:56Z
ethz.source
SCOPUS
ethz.eth
yes
en_US
ethz.availability
Metadata only
en_US
ethz.rosetta.installDate
2021-09-30T07:45:19Z
ethz.rosetta.lastUpdated
2022-03-29T13:41:29Z
ethz.rosetta.versionExported
true
ethz.COinS
ctx_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.atitle=A%20Microcontroller%20is%20All%20You%20Need:%20Enabling%20Transformer%20Execution%20on%20Low-Power%20IoT%20Endnodes&rft.date=2021&rft.spage=9524173&rft.au=Burrello,%20Alessio&Scherer,%20Moritz&Zanghieri,%20Marcello&Conti,%20Francesco&Benini,%20Luca&rft.isbn=978-1-6654-3156-9&978-1-6654-3157-6&rft.genre=proceeding&rft_id=info:doi/10.1109/COINS51742.2021.9524173&rft.btitle=2021%20IEEE%20International%20Conference%20on%20Omni-Layer%20Intelligent%20Systems%20(COINS)
 Printexemplar via ETH-Bibliothek suchen

Dateien zu diesem Eintrag

DateienGrößeFormatIm Viewer öffnen

Zu diesem Eintrag gibt es keine Dateien.

Publikationstyp

Zur Kurzanzeige