Bayesian Sparsification for Deep Neural Networks With Bayesian Model Reduction
Research output: Contribution to journal › Research article › Contributed › peer-review
Contributors
Abstract
Deep learning's immense capabilities are often constrained by the complexity of its models, leading to an increasing demand for effective sparsification techniques. Bayesian sparsification for deep learning emerges as a crucial approach, facilitating the design of models that are both computationally efficient and competitive in terms of performance across various deep learning applications. The state-of-the-art - in Bayesian sparsification of deep neural networks - combines structural shrinkage priors on model weights with an approximate inference scheme based on stochastic variational inference. However, model inversion of the full generative model is exceptionally computationally demanding, especially when compared to standard deep learning of point estimates. In this context, we advocate for the use of Bayesian model reduction (BMR) as a more efficient alternative for pruning of model weights. As a generalization of the Savage-Dickey ratio, BMR allows a post-hoc elimination of redundant model weights based on the posterior estimates under a straightforward (non-hierarchical) generative model. Our comparative study highlights the advantages of the BMR method relative to established approaches, which are based on hierarchical horseshoe priors over model weights. We illustrate the potential of BMR across various deep learning architectures, from classical networks like LeNet to modern frameworks such as Vision Transformers and MLP-Mixers.
Details
Original language | English |
---|---|
Pages (from-to) | 88231-88242 |
Number of pages | 12 |
Journal | IEEE access |
Volume | 12 (2024) |
Publication status | Published - 20 Jun 2024 |
Peer-reviewed | Yes |
Keywords
ASJC Scopus subject areas
Keywords
- Bayesian model reduction, deep neural networks, stochastic variational inference