Hardware implementation of backpropagation using progressive gradient descent for in situ training of multilayer neural networks

Research output: Contribution to journalArticleAcademicpeer-review

11 Citations (Scopus)
121 Downloads (Pure)

Abstract

Neural network training can be slow and energy-expensive due to the frequent transfer of weight data between digital memory and processing units. Neuromorphic systems can accelerate neural networks by performing multiply-accumulate operations in parallel using nonvolatile analog memory. However, executing the widely used backpropagation training algorithm in multilayer neural networks requires information-and therefore storage-of the partial derivatives of the weight values preventing suitable and scalable implementation in hardware. Here, we propose a hardware implementation of the backpropagation algorithm that progressively updates each layer using in situ stochastic gradient descent, avoiding this storage requirement. We experimentally demonstrate the in situ error calculation and the proposed progressive backpropagation method in a multilayer hardware-implemented neural network. We confirm identical learning characteristics and classification performance compared to conventional backpropagation in software. We show that our approach can be scaled to large and deep neural networks, enabling highly efficient training of advanced artificial intelligence computing systems.

Original languageEnglish
Article numbereado8999
Number of pages7
JournalScience Advances
Volume10
Issue number28
DOIs
Publication statusPublished - Jul 2024

Funding

FundersFunder number
European Union's Horizon 2020 - Research and Innovation Framework Programme
Eindhoven Hendrik Casimir institute
Eindhoven Hendrik Casimir institute
European Union's Horizon 2020 - Research and Innovation Framework Programme802615

    Fingerprint

    Dive into the research topics of 'Hardware implementation of backpropagation using progressive gradient descent for in situ training of multilayer neural networks'. Together they form a unique fingerprint.

    Cite this