JMLR

Posterior and Variational Inference for Deep Neural Networks with Heavy-Tailed Weights

Authors
Paul Egels Ismaël Castillo
Research Topics
Machine Learning Bayesian Statistics
Paper Information
  • Journal:
    Journal of Machine Learning Research
  • Added to Tracker:
    Sep 08, 2025
Abstract

We consider deep neural networks in a Bayesian framework with a prior distribution sampling the network weights at random. Following a recent idea of Agapiou and Castillo (2024), who show that heavy-tailed prior distributions achieve automatic adaptation to smoothness, we introduce a simple Bayesian deep learning prior based on heavy-tailed weights and ReLU activation. We show that the corresponding posterior distribution achieves near-optimal minimax contraction rates, simultaneously adaptive to both intrinsic dimension and smoothness of the underlying function, in a variety of contexts including nonparametric regression, geometric data and Besov spaces. While most works so far need a form of model selection built-in within the prior distribution, a key aspect of our approach is that it does not require to sample hyperparameters to learn the architecture of the network. We also provide variational Bayes counterparts of the results, that show that mean-field variational approximations still benefit from near-optimal theoretical support.

Author Details
Paul Egels
Author
Ismaël Castillo
Author
Research Topics & Keywords
Machine Learning
Research Area
Bayesian Statistics
Research Area
Citation Information
APA Format
Paul Egels & Ismaël Castillo . Posterior and Variational Inference for Deep Neural Networks with Heavy-Tailed Weights. Journal of Machine Learning Research .
BibTeX Format
@article{paper533,
  title = { Posterior and Variational Inference for Deep Neural Networks with Heavy-Tailed Weights },
  author = { Paul Egels and Ismaël Castillo },
  journal = { Journal of Machine Learning Research },
  url = { https://www.jmlr.org/papers/v26/24-0894.html }
}