Samenvatting
Edge devices can execute pre-trained Artificial Intelligence (AI) models optimized on large Graphical Processing Units (GPU) but often need fine-tuning for real-world data. This process, known as edge learning, is crucial for personalized learning for tasks such as speech and gesture recognition and often requires recurrent neural networks (RNNs). However, training RNNs on edge devices faces challenges due to limited resources. We propose a system for RNN training through sequence partitioning using the Forward Propagation Through Time (FPTT) training method, facilitating edge learning. Our optimized HW/SW co-design for FPTT is the first of its kind. In our work, we have implemented the complete computational process for training Long Short-Term Memory (LSTM) networks using FPTT, and we have optimized and explored the hardware architecture leveraging the Chipyard framework. Our findings indicate considerable memory savings, with only a slight increase in latency, when training small-batch size sequential MNIST (S-MNIST) data.
Originele taal-2 | Engels |
---|---|
Titel | 2024 IFIP/IEEE 32nd International Conference on Very Large Scale Integration, VLSI-SoC 2024 |
Uitgeverij | Institute of Electrical and Electronics Engineers |
Aantal pagina's | 4 |
ISBN van elektronische versie | 979-8-3315-3967-2 |
DOI's | |
Status | Gepubliceerd - 3 dec. 2024 |
Evenement | IFIP/IEEE International Conference on Very Large Scale Integration, VLSI-SoC 2024 - Morocco, Tanger, Marokko Duur: 6 okt. 2024 → 9 okt. 2024 https://vlsisoc2024.nl/ |
Congres
Congres | IFIP/IEEE International Conference on Very Large Scale Integration, VLSI-SoC 2024 |
---|---|
Verkorte titel | VLSI-SoC 2024 |
Land/Regio | Marokko |
Stad | Tanger |
Periode | 6/10/24 → 9/10/24 |
Internet adres |
Financiering
This work has been funded by the Dutch Organization for Scientific Research (NWO) with Grant KICH1.ST04.22.021.
Financiers | Financiernummer |
---|---|
Netherlands Organisation for Applied Scientific Research - TNO | KICH1.ST04.22.021 |