TY - GEN
T1 - ELSE
T2 - 18th European Conference on Computer Vision, ECCV 2024
AU - Zhu, Zeqi
AU - Garcia-Ortiz, Alberto
AU - Waeijen, Luc
AU - Bondarev, Egor
AU - Pourtaherian, Arash
AU - Moreira, Orlando
PY - 2024/11/1
Y1 - 2024/11/1
N2 - Brain-inspired computer architecture facilitates low-power, low-latency deep neural network inference for embedded AI applications. The hardware performance crucially hinges on the quantity of non-zero activations (i.e., events) during inference. Thus, we propose a novel event suppression method, dubbed ELSE, which enhances inference Efficiency via Line-based Sparsity Exploration. Specifically, it exploits spatial correlation between adjacent lines in activation maps to reduce network events. ELSE reduces event-triggered computations by 3.14–6.49× for object detection and by 2.43–5.75× for pose estimation across various network architectures compared to conventional processing. Additionally, we show that combining ELSE with other event suppression methods can either significantly enhance computation savings for spatial suppression or reduce state memory footprint by >2× for temporal suppression. The latter alleviates the challenge of temporal execution exceeding the resource constraints of real-world embedded platforms. These results highlight ELSE’s significant event suppression ability and its capacity to deliver complementary performance enhancements for SOTA methods.
AB - Brain-inspired computer architecture facilitates low-power, low-latency deep neural network inference for embedded AI applications. The hardware performance crucially hinges on the quantity of non-zero activations (i.e., events) during inference. Thus, we propose a novel event suppression method, dubbed ELSE, which enhances inference Efficiency via Line-based Sparsity Exploration. Specifically, it exploits spatial correlation between adjacent lines in activation maps to reduce network events. ELSE reduces event-triggered computations by 3.14–6.49× for object detection and by 2.43–5.75× for pose estimation across various network architectures compared to conventional processing. Additionally, we show that combining ELSE with other event suppression methods can either significantly enhance computation savings for spatial suppression or reduce state memory footprint by >2× for temporal suppression. The latter alleviates the challenge of temporal execution exceeding the resource constraints of real-world embedded platforms. These results highlight ELSE’s significant event suppression ability and its capacity to deliver complementary performance enhancements for SOTA methods.
KW - activation sparsity exploration
KW - efficient neural networks
UR - http://www.scopus.com/inward/record.url?scp=85210012634&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-73247-8_24
DO - 10.1007/978-3-031-73247-8_24
M3 - Conference contribution
AN - SCOPUS:85210012634
SN - 978-3-031-73246-1
T3 - Lecture Notes in Computer Science (LNCS)
SP - 412
EP - 431
BT - Computer Vision – ECCV 2024
A2 - Leonardis, Aleš
A2 - Ricci, Elisa
A2 - Roth, Stefan
A2 - Russakovsky, Olga
A2 - Sattler, Torsten
A2 - Varol, Gül
PB - Springer
CY - Cham
Y2 - 29 September 2024 through 4 October 2024
ER -