Assumed audience:
ML people
“Gaussian Processes” are stochastic processes/fields with joint Gaussian distributions over all finite sets of observation locations. The most familiar to finance and physics people is the Gauss-Markov process, a.k.a. the Wiener process, but there are many others. These processes are convenient due to certain useful properties of the multivariate Gaussian distribution, e.g., being uniquely specified by first and second moments, nice behaviour under various linear operations, and kernel tricks. Especially famous applications include Gaussian process regression and spatial statistics. Check out Ti’s Interactive visualization for some examples.
Gaussian processes are, specifically, probabilistic distributions over random functions for some index (or argument) set , often taken to be .
We typically work with a mean-zero process, meaning for every finite set of observations of that process, the joint distribution is mean-zero Gaussian, where is the sample covariance matrix defined such that its entries are given by This is the covariance kernel that maps from function argument — — to the second moment of function values. In this case, we are specifying only the second moments, which gives all the remaining properties of the process.
Derivatives and integrals
Derivative of a Gaussian process
TBD.
For now, see these blog posts:
I am using results from Adler (2010), Adler and Taylor (2007). See also pathwise GPs for some useful results here.
References
Adler. 2010. The Geometry of Random Fields.
Adler, and Taylor. 2007.
Random Fields and Geometry. Springer Monographs in Mathematics 115.
Agrell. n.d. “Gaussian Processes with Linear Operator Inequality Constraints.”
Alexanderian. 2015.
“A Brief Note on the Karhunen-Loève Expansion.” arXiv:1509.07526 [Math].
Bochner. 1959. Lectures on Fourier Integrals.
Dym, and McKean. 2008. Gaussian Processes, Function Theory, and the Inverse Spectral Problem. Dover Books on Mathematics.
Kanagawa, and Fukumizu. 2014.
“Recovering Distributions from Gaussian RKHS Embeddings.” In
Journal of Machine Learning Research.
Lange-Hegermann. 2018.
“Algorithmic Linearly Constrained Gaussian Processes.” In
Proceedings of the 32nd International Conference on Neural Information Processing Systems. NIPS’18.
———. 2021.
“Linearly Constrained Gaussian Processes with Boundary Conditions.” In
Proceedings of the 24th International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research.
Liu, and Röckner. 2015. Stochastic Partial Differential Equations: An Introduction.
Long, Wang, Krishnapriyan, et al. 2022.
“AutoIP: A United Framework to Integrate Physics into Gaussian Processes.” In
Proceedings of the 39th International Conference on Machine Learning.
Lukić, and Beder. 2001.
“Stochastic Processes with Sample Paths in Reproducing Kernel Hilbert Spaces.” Transactions of the American Mathematical Society.
Rasmussen, and Nickisch. 2010.
“Gaussian Processes for Machine Learning (GPML) Toolbox.” Journal of Machine Learning Research.
Rasmussen, and Williams. 2006.
Gaussian Processes for Machine Learning. Adaptive Computation and Machine Learning.
Taylor. 2009.
“Random Fields.”
Yaglom. 1987. Correlation Theory of Stationary and Related Random Functions. Volume II: Supplementary Notes and References. Springer Series in Statistics.