TY - JOUR
T1 - Sparse evolutionary deep learning with over one million artificial neurons on commodity hardware
AU - Liu, Shiwei
AU - Mocanu, Decebal
AU - Ramapuram Matavalam, Amarsagar Reddy
AU - Pei, Yulong
AU - Pechenizkiy, Mykola
PY - 2021/4
Y1 - 2021/4
N2 - Artificial Neural Networks (ANNs) have emerged as hot topics in the research community. Despite the success of ANNs, it is challenging to train and deploy modern ANNs on commodity hardware due to the ever-increasing model size and the unprecedented growth in the data volumes. Particularly for microarray data, the very-high dimensionality and the small number of samples make it difficult for machine learning techniques to handle. Furthermore, specialized hardware such as Graphics Processing Unit (GPU) is expensive. Sparse neural networks are the leading approaches to address these challenges. However, off-the-shelf sparsity inducing techniques either operate from a pre-trained model or enforce the sparse structure via binary masks. The training efficiency of sparse neural networks cannot be obtained practically. In this paper, we introduce a technique allowing us to train truly sparse neural networks with fixed parameter count throughout training. Our experimental results demonstrate that our method can be applied directly to handle high dimensional data, while achieving higher accuracy than the traditional two phases approaches. Moreover, we have been able to create truly sparse MultiLayer Perceptrons (MLPs) models with over one million neurons and to train them on a typical laptop without GPU ( \url{https://github.com/dcmocanu/sparse-evolutionary-artificial-neural-networks/tree/master/SET-MLP-Sparse-Python-Data-Structures}), this being way beyond what is possible with any state-of-the-art technique.
AB - Artificial Neural Networks (ANNs) have emerged as hot topics in the research community. Despite the success of ANNs, it is challenging to train and deploy modern ANNs on commodity hardware due to the ever-increasing model size and the unprecedented growth in the data volumes. Particularly for microarray data, the very-high dimensionality and the small number of samples make it difficult for machine learning techniques to handle. Furthermore, specialized hardware such as Graphics Processing Unit (GPU) is expensive. Sparse neural networks are the leading approaches to address these challenges. However, off-the-shelf sparsity inducing techniques either operate from a pre-trained model or enforce the sparse structure via binary masks. The training efficiency of sparse neural networks cannot be obtained practically. In this paper, we introduce a technique allowing us to train truly sparse neural networks with fixed parameter count throughout training. Our experimental results demonstrate that our method can be applied directly to handle high dimensional data, while achieving higher accuracy than the traditional two phases approaches. Moreover, we have been able to create truly sparse MultiLayer Perceptrons (MLPs) models with over one million neurons and to train them on a typical laptop without GPU ( \url{https://github.com/dcmocanu/sparse-evolutionary-artificial-neural-networks/tree/master/SET-MLP-Sparse-Python-Data-Structures}), this being way beyond what is possible with any state-of-the-art technique.
KW - scalable deep learning
KW - sparse neural network
KW - microarray gene expression
KW - Microarray gene expression
KW - Adaptive sparse connectivity
KW - Sparse evolutionary training (SET)
KW - Truly sparse neural networks
UR - http://www.scopus.com/inward/record.url?scp=85087634918&partnerID=8YFLogxK
U2 - 10.1007/s00521-020-05136-7
DO - 10.1007/s00521-020-05136-7
M3 - Article
SN - 0941-0643
VL - 33
SP - 2589
EP - 2604
JO - Neural Computing and Applications
JF - Neural Computing and Applications
IS - 7
ER -