AEBNAS: Strengthening Exit Branches in Early-Exit Networks through Hardware-Aware Neural Architecture Search

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

Abstract

Early-exit networks are effective solutions for reducing the overall energy consumption and latency of deep learning models by adjusting computation based on the complexity of input data. By incorporating intermediate exit branches into the architecture, they provide less computation for simpler samples, which is particularly beneficial for resource-constrained devices where energy consumption is crucial. However, designing early-exit networks is a challenging and time-consuming process due to the need to balance efficiency and performance. Recent works have utilized Neural Architecture Search (NAS) to design more efficient early-exit networks, aiming to reduce average latency while improving model accuracy by determining the best positions and number of exit branches in the architecture. Another important factor affecting the efficiency and accuracy of early-exit networks is the depth and types of layers in the exit branches. In this paper, we use hardware-aware NAS to strengthen exit branches, considering both accuracy and efficiency during optimization. Our performance evaluation on the CIFAR-10, CIFAR-100, and SVHN datasets demonstrates that our proposed framework, which considers varying depths and layers for exit branches along with adaptive threshold tuning, designs early-exit networks that achieve higher accuracy with the same or lower average number of MACs compared to the state-of-the-art approaches.
Original languageEnglish
Title of host publication2025 3rd IEEE International Conference on Federated Learning Technologies and Applications, FLTA
PublisherInstitute of Electrical and Electronics Engineers
Number of pages8
ISBN (Electronic)979-8-3315-5670-9
DOIs
Publication statusPublished - 20 Jan 2026
Event3rd International Conference on Federated Learning Technologies and Applications, FLTA - Dubrovnik, Croatia
Duration: 14 Oct 202517 Oct 2025

Conference

Conference3rd International Conference on Federated Learning Technologies and Applications, FLTA
Abbreviated titleFLTA
Country/TerritoryCroatia
CityDubrovnik
Period14/10/2517/10/25

UN SDGs

This output contributes to the following UN Sustainable Development Goals (SDGs)

  1. SDG 7 - Affordable and Clean Energy
    SDG 7 Affordable and Clean Energy

Keywords

  • Early-Exit Networks
  • Hardware-Aware NAS
  • Efficient Deep Learning, Dynamic Neural Networks

Fingerprint

Dive into the research topics of 'AEBNAS: Strengthening Exit Branches in Early-Exit Networks through Hardware-Aware Neural Architecture Search'. Together they form a unique fingerprint.

Cite this