Show simple item record

dc.contributor.author
Rutishauser, Georg
dc.contributor.author
Mihali, Joan
dc.contributor.author
Scherer, Moritz
dc.contributor.author
Bonini, Lucas
dc.date.accessioned
2024-10-17T10:00:35Z
dc.date.available
2024-10-17T06:00:51Z
dc.date.available
2024-10-17T10:00:35Z
dc.date.issued
2024
dc.identifier.isbn
979-8-3503-4963-4
en_US
dc.identifier.isbn
979-8-3503-4964-1
en_US
dc.identifier.issn
2160-0511
dc.identifier.other
10.1109/ASAP61560.2024.00049
en_US
dc.identifier.uri
http://hdl.handle.net/20.500.11850/700163
dc.description.abstract
Ternary neural networks (TNNs) offer a superior accuracy-energy trade-off compared to binary neural networks. However, until now, they have required specialized accelerators to realize their efficiency potential, which has hindered widespread adoption. To address this, we present xTern, a lightweight extension of the RISC-V instruction set architecture (ISA) targeted at accelerating TNN inference on general-purpose cores. To complement the ISA extension, we developed a set of optimized kernels leveraging xTern, achieving 67% higher throughput than their 2-bit equivalents. Power consumption is only marginally increased by 5.2 %, resulting in an energy efficiency improvement by 57.1 %. We demonstrate that the proposed xTern extension, integrated into an octa-core compute cluster, incurs a minimal silicon area overhead of 0.9% with no impact on timing. In end-to-end benchmarks, we demonstrate that xTern enables the deployment of TNNs achieving up to 1.6 percentage points higher CIFAR-10 classification accuracy than 2-bit networks at equal inference latency. Our results show that xTern enables RISCV-based ultra-low-power edge AI platforms to benefit from the efficiency potential of TNNs.
en_US
dc.language.iso
en
en_US
dc.publisher
IEEE
en_US
dc.title
xTern: Energy-Efficient Ternary Neural Network Inference on RISC-V-Based Edge Systems
en_US
dc.type
Conference Paper
ethz.book.title
2024 IEEE 35th International Conference on Application-specific Systems, Architectures and Processors (ASAP)
en_US
ethz.pages.start
206
en_US
ethz.pages.end
213
en_US
ethz.event
35th IEEE International Conference on Application-Specific Systems, Architectures and Processors (ASAP 2024)
en_US
ethz.event.location
Hong Kong, Hong Kong
en_US
ethz.event.date
July 24-26, 2024
en_US
ethz.notes
Conference Presentation held on July 24, 2024.
en_US
ethz.grant
A multiprocessor system on chip with in-memory neural processing unit
en_US
ethz.identifier.wos
ethz.publication.place
Piscataway, NJ
en_US
ethz.publication.status
published
en_US
ethz.grant.agreementno
101070634
ethz.grant.fundername
SBFI
ethz.grant.funderDoi
10.13039/501100007352
ethz.grant.program
HE
ethz.date.deposited
2024-10-17T06:01:02Z
ethz.source
WOS
ethz.eth
yes
en_US
ethz.availability
Metadata only
en_US
ethz.rosetta.installDate
2024-10-17T10:00:36Z
ethz.rosetta.lastUpdated
2024-10-17T10:00:36Z
ethz.rosetta.versionExported
true
ethz.COinS
ctx_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.atitle=xTern:%20Energy-Efficient%20Ternary%20Neural%20Network%20Inference%20on%20RISC-V-Based%20Edge%20Systems&rft.date=2024&rft.spage=206&rft.epage=213&rft.issn=2160-0511&rft.au=Rutishauser,%20Georg&Mihali,%20Joan&Scherer,%20Moritz&Bonini,%20Lucas&rft.isbn=979-8-3503-4963-4&979-8-3503-4964-1&rft.genre=proceeding&rft_id=info:doi/10.1109/ASAP61560.2024.00049&rft.btitle=2024%20IEEE%2035th%20International%20Conference%20on%20Application-specific%20Systems,%20Architectures%20and%20Processors%20(ASAP)
 Search print copy at ETH Library

Files in this item

FilesSizeFormatOpen in viewer

There are no files associated with this item.

Publication type

Show simple item record