There are many ways to cleverly slice up GP likelihoods so that inference is cheap.
This page is about some of them, especially the union of sparse and variational tricks. Scalable Gaussian process regressions choose cunning factorisations such that the model collapses down to a lower-dimensional thing than it might have seemed to need, at least approximately. There is a compilation of tricks to make this go — variational approximations a model, sparse GP models where there are a small number of inducing points (Dezfouli and Bonilla 2015; Edwin V. Bonilla, Krauth, and Dezfouli 2019; Krauth et al. 2016; Hensman, Fusi, and Lawrence 2013; Salimbeni and Deisenroth 2017). You might suspect yourself of using such a method if you find that some important high-dimensional expectation can be evaluated by some function of univariate Gaussians.
There are indeed a lot of different factorisations that can be done here; See filtering GPs for one which factorises over a single input axis. Other tricks from outside GPs which factorise a distribution cleverly, such as message-passing inference are likely useful here. Also, Toeplitz and related structures work out nicely for, e.g. lattice-distributed inputs and some other situations I forget right now.
SVI for Gaussian processes
As seen in Hensman, Fusi, and Lawrence (2013);Salimbeni and Deisenroth (2017).
Vecchia factorisation
Approximate the precision matrix by one with a sparse Cholesky factorisation. See Vecchia factorisation.
Latent Gaussian Process models
The Edwin V. Bonilla, Krauth, and Dezfouli (2019) set up for Latent Gaussian Process models (“LGPMs”) goes as follows:
We are learning a mapping from data. The dataset looks like is an input vector and is an output. We decree that the mapping from inputs to outputs may be expressed by underlying latent functions We assume that the latent functions are drawn from (a priori) independent zero-mean Gaussian processes.
Here is the set of all latent function values; denotes the values of latent function . The Gram matrix is , induced by a covariance kernel, . The parameters of all kernel functions we call Our observation model can have various likelihoods; We call the corresponding parameter . We assume that our multi-dimensional observations are i.i.d. given the latent functions so that is the set of latent values upon which depends.
There are several factorizations to note here
- The prior is factored into latent functions per-coordinate
- the conditional likelihood is factored over observations (i.e. noise is independent)
If we further factorise the variational approximation in some way this will work out nicely, e.g. into Gaussian mixtures. This works out well for us when we try to devise a system of inference later to minimise the ELBO. TBC.
For now, though, let us examine exactly tractable inference
References
Adam, Eleftheriadis, Durrande, et al. 2020.
“Doubly Sparse Variational Gaussian Processes.” In
AISTATS.
Bonilla, Edwin V. 2017. “Variational Learning of GP Models.”
Bonilla, Edwin V., Chai, and Williams. 2007.
“Multi-Task Gaussian Process Prediction.” In
Proceedings of the 20th International Conference on Neural Information Processing Systems. NIPS’07.
Bonilla, Edwin V., Krauth, and Dezfouli. 2019.
“Generic Inference in Latent Gaussian Process Models.” Journal of Machine Learning Research.
Bruinsma, Perim, Tebbutt, et al. 2020.
“Scalable Exact Inference in Multi-Output Gaussian Processes.” In
International Conference on Machine Learning.
Dahl, and Bonilla. 2017.
“Scalable Gaussian Process Models for Solar Power Forecasting.” In
Data Analytics for Renewable Energy Integration: Informing the Generation and Distribution of Renewable Energy. Lecture Notes in Computer Science.
Dezfouli, and Bonilla. 2015.
“Scalable Inference for Gaussian Process Models with Black-Box Likelihoods.” In
Advances in Neural Information Processing Systems 28. NIPS’15.
Dutordoir, Hensman, van der Wilk, et al. 2021.
“Deep Neural Networks as Point Estimates for Deep Gaussian Processes.” In
arXiv:2105.04504 [Cs, Stat].
Hensman, Fusi, and Lawrence. 2013.
“Gaussian Processes for Big Data.” In
Proceedings of the Twenty-Ninth Conference on Uncertainty in Artificial Intelligence. UAI’13.
Lázaro-Gredilla, and Figueiras-Vidal. 2009.
“Inter-Domain Gaussian Processes for Sparse Inference Using Inducing Features.” In
Advances in Neural Information Processing Systems.
Lemercier, Salvi, Cass, et al. 2021.
“SigGPDE: Scaling Sparse Gaussian Processes on Sequential Data.” In
Proceedings of the 38th International Conference on Machine Learning.
Meanti, Carratino, Rosasco, et al. 2020.
“Kernel Methods Through the Roof: Handling Billions of Points Efficiently.” In
Proceedings of the 34th International Conference on Neural Information Processing Systems. NIPS’20.
Nguyen, and Bonilla. 2014.
“Automated Variational Inference for Gaussian Process Models.” In
Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 1. NIPS’14.
Qi, Abdel-Gawad, and Minka. 2010.
“Sparse-Posterior Gaussian Processes for General Likelihoods.” In
Proceedings of the Twenty-Sixth Conference on Uncertainty in Artificial Intelligence. UAI’10.
Rossi, Heinonen, Bonilla, et al. 2021.
“Sparse Gaussian Processes Revisited: Bayesian Approaches to Inducing-Variable Approximations.” In
Proceedings of The 24th International Conference on Artificial Intelligence and Statistics.
Saatçi, Turner, and Rasmussen. 2010.
“Gaussian Process Change Point Models.” In
Proceedings of the 27th International Conference on International Conference on Machine Learning. ICML’10.
Salimbeni, and Deisenroth. 2017.
“Doubly Stochastic Variational Inference for Deep Gaussian Processes.” In
Advances In Neural Information Processing Systems.
Shi, Titsias, and Mnih. 2020.
“Sparse Orthogonal Variational Inference for Gaussian Processes.” In
Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics.
Snelson, and Ghahramani. 2007.
“Local and Global Sparse Gaussian Process Approximations.” In
Proceedings of the Eleventh International Conference on Artificial Intelligence and Statistics.
Titsias. 2009.
“Variational Learning of Inducing Variables in Sparse Gaussian Processes.” In
International Conference on Artificial Intelligence and Statistics.
Wilson, Knowles, and Ghahramani. 2012.
“Gaussian Process Regression Networks.” In
Proceedings of the 29th International Coference on International Conference on Machine Learning. ICML’12.
Zhang, Rui, Walder, Bonilla, et al. 2020.
“Quantile Propagation for Wasserstein-Approximate Gaussian Processes.” In
Proceedings of NeurIPS 2020.