Finite Neural Networks as Mixtures of Gaussian Processes: From Provable Error Bounds to Prior Selection
Authors
Research Topics
Paper Information
-
Journal:
Journal of Machine Learning Research -
Added to Tracker:
Mar 03, 2026
Abstract
Infinitely wide or deep neural networks (NNs) with independent and identically distributed (i.i.d.) parameters have been shown to be equivalent to Gaussian processes. Because of the favorable properties of Gaussian processes, this equivalence is commonly employed to analyze neural networks and has led to various breakthroughs over the years. However, neural networks and Gaussian processes are equivalent only in the limit; in the finite case there are currently no methods available to approximate a trained neural network with a Gaussian model with bounds on the approximation error. In this work, we present an algorithmic framework to approximate a neural network of finite width and depth, and with not necessarily i.i.d. parameters, with a mixture of Gaussian processes with bounds on the approximation error. In particular, we consider the Wasserstein distance to quantify the closeness between probabilistic models and, by relying on tools from optimal transport and Gaussian processes, we iteratively approximate the output distribution of each layer of the neural network as a mixture of Gaussian processes. Crucially, for any NN and $\epsilon >0$ our approach is able to return a mixture of Gaussian processes that is $\epsilon$-close to the NN at a finite set of input points. Furthermore, we rely on the differentiability of the resulting error bound to show how our approach can be employed to tune the parameters of a NN to mimic the functional behavior of a given Gaussian process, e.g., for prior selection in the context of Bayesian inference. We empirically investigate the effectiveness of our results on both regression and classification problems with various neural network architectures. Our experiments highlight how our results can represent an important step towards understanding neural network predictions and formally quantifying their uncertainty.
Author Details
Steven Adams
AuthorAndrea Patanè
AuthorMorteza Lahijanian
AuthorLuca Laurenti
AuthorResearch Topics & Keywords
Machine Learning
Research AreaBayesian Statistics
Research AreaCitation Information
APA Format
Steven Adams
,
Andrea Patanè
,
Morteza Lahijanian
&
Luca Laurenti
.
Finite Neural Networks as Mixtures of Gaussian Processes: From Provable Error Bounds to Prior Selection.
Journal of Machine Learning Research
.
BibTeX Format
@article{paper981,
title = { Finite Neural Networks as Mixtures of Gaussian Processes: From Provable Error Bounds to Prior Selection },
author = {
Steven Adams
and Andrea Patanè
and Morteza Lahijanian
and Luca Laurenti
},
journal = { Journal of Machine Learning Research },
url = { https://www.jmlr.org/papers/v27/24-1199.html }
}