Sparse Training of Neural Networks based on Multilevel Mirror Descent
Abstract
We introduce a dynamic sparse training algorithm based on linearized Bregman iterations / mirror descent that exploits the naturally incurred sparsity by alternating between periods of static and dynamic sparsity pattern updates. The key idea is to combine sparsity-inducing Bregman iterations with adaptive freezing of the network structure to enable efficient exploration of the sparse parameter space while maintaining sparsity. We provide convergence guaranties by embedding our method in a multilevel optimization framework. Furthermore, we empirically show that our algorithm can produce highly sparse and accurate models on standard benchmarks. We also show that the theoretical number of FLOPs compared to SGD training can be reduced from 38% for standard Bregman iterations to 6% for our method while maintaining test accuracy.
Growth and citations
This paper is currently showing No growth state computed yet..
Citation metrics and growth state from academic sources (e.g. Semantic Scholar). See About for details.
Cited by (0)
No citing papers yet
Papers that cite this one will appear here once data is available.
View citations page →References (0)
No references in DB yet
References for this paper will appear here once ingested.
Related papers in Numerical Analysis
- Noisy nonlocal aggregation model with gradient flow structures0 citations
- Do We Need Asynchronous SGD? On the Near-Optimality of Synchronous Methods0 citations
- Reducing acquisition time and radiation damage: data-driven subsampling for spectro-microscopy0 citations
Growth transitions
No transitions recorded yet
Growth state transitions will appear here once computed.