BrainTTA: A 28.6 TOPS/W Compiler Programmable Transport Triggered NN SoC

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

3 Citations (Scopus)
11 Downloads (Pure)

Abstract

Accelerators designed for deep neural network (DNN) inference with extremely low operand widths, down to 1-bit, have become popular due to their ability to significantly reduce energy consumption during inference. This paper introduces a compiler-programmable flexible System-on-Chip (SoC) with mixed-precision support. This SoC is based on a Transport-Triggered Architecture (TTA) that facilitates efficient implementation of DNN workloads. By shifting the complexity of data movement from the hardware scheduler to the exposed-datapath compiler, DNN workloads can be implemented in an energy efficient yet flexible way. The architecture is fully supported by a compiler and can be programmed using C/C++/OpenCL. The SoC is implemented using 22nm FDX technology and achieves a peak energy efficiency of 28.6/14.9/2.47 TOPS/W for binary, ternary, and 8-bit precision, respectively, while delivering a throughput of 614/307/77 GOPS. Compared to state-of-the-art (SotA), this work achieves up to 3.3x better energy efficiency compared to other programmable solutions.
Original languageEnglish
Title of host publication2023 IEEE International Conference on Computer Design (ICCD)
PublisherInstitute of Electrical and Electronics Engineers
Pages78-85
Number of pages8
ISBN (Electronic)979-8-3503-4291-8
DOIs
Publication statusPublished - 22 Dec 2023
Event41st IEEE International Conference on Computer Design, ICCD 2023 - Washington, United States
Duration: 6 Nov 20238 Nov 2023

Conference

Conference41st IEEE International Conference on Computer Design, ICCD 2023
Abbreviated titleICCD 2023
Country/TerritoryUnited States
CityWashington
Period6/11/238/11/23

Keywords

  • Binary neural network
  • Ternary neural network
  • deep learning
  • edge computing
  • low power inference
  • mixed precision accelerator
  • neural network accelerator

Fingerprint

Dive into the research topics of 'BrainTTA: A 28.6 TOPS/W Compiler Programmable Transport Triggered NN SoC'. Together they form a unique fingerprint.

Cite this