On improving deep learning generalization with adaptive sparse connectivity

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

18 Downloads (Pure)


Large neural networks are very successful in various tasks. However, with limited data, the generalization capabilities of deep neural networks are also very limited. In this paper, we empirically start showing that intrinsically sparse neural networks with adaptive sparse connectivity, which by design have a strict parameter budget during the training phase, have better generalization capabilities than their fully-connected counterparts. Besides this, we propose a new technique to train these sparse models by combining the Sparse Evolutionary Training (SET) procedure with neurons pruning. Operated on MultiLayer Perceptron (MLP) and tested on 15 datasets, our proposed technique zeros out around 50% of the hidden neurons during training, while having a linear number of parameters to optimize with respect to the number of neurons. The results show a competitive classification and generalization performance.
Original languageEnglish
Title of host publicationICML 2019 Workshop on Understanding and Improving General-ization in Deep Learning
Number of pages5
Publication statusPublished - 14 Jun 2019


  • adaptive sparse connectivity
  • eep learning generalization
  • neurons prunning
  • sparse evolutionary training
  • neural networks


Dive into the research topics of 'On improving deep learning generalization with adaptive sparse connectivity'. Together they form a unique fingerprint.

Cite this