Handling and computing probability measures via reproducing kernel methods. If we would like to make a dependence test or a probability metric from this trick then we famously can; that is the integral probability metric that we call maximum mean discrepancy.
Operations
From Kernel embedding of distributions on Wikipedia:
As the Wikipedia comments observe, the notation is unusual; It only really makes sense in the context of the paper.
This section illustrates how basic probabilistic rules may be reformulated as (multi)linear algebraic operations in the kernel embedding framework and is primarily based on the work of Song et al (Song et al. 2009; Song, Fukumizu, and Gretton 2013). The following notation is adopted:
- joint distribution over random variables
- marginal distribution of marginal distribution of
- conditional distribution of given with corresponding conditional embedding operator
- prior distribution over
- is used to distinguish distributions which incorporate the prior from distributions which do not rely on the prior
In practice, all embeddings are empirically estimated from data and it is assumed that a set of samples may be used to estimate the kernel embedding of the prior distribution .
Kernel sum rule
In probability theory, the marginal distribution of can be computed by integrating out from the joint density (including the prior distribution on )
The analogue of this rule in the kernel embedding framework states that , the RKHS embedding of , can be computed via where is the kernel embedding of . In practical implementations, the kernel sum rule takes the following form where is the empirical kernel embedding of the prior distribution, , and are Gram matrices with entries respectively.
Kernel chain rule
In probability theory, a joint distribution can be factorised into a product between conditional and marginal distributions
The analogue of this rule in the kernel embedding framework states that , the joint embedding of , can be factorised as a composition conditional embedding operator with the auto-covariance operator associated with where
In practical implementations, the kernel chain rule takes the following form
Kernel Bayes’ rule
In probability theory, a posterior distribution can be expressed in terms of a prior distribution and a likelihood function as
The analogue of this rule in the kernel embedding framework expresses the kernel embedding of the conditional distribution in terms of conditional embedding operators which are modified by the prior distribution where from the chain rule:
In practical implementations, the kernel Bayes’ rule takes the following form where
Two regularisation parameters are used in this framework: for the estimation of and for the estimation of the final conditional embedding operator## References
The latter regularisation is done on square of because may not be positive definite.
References
Cherief-Abdellatif, and Alquier. 2020.
“MMD-Bayes: Robust Bayesian Estimation via Maximum Mean Discrepancy.” In
Proceedings of The 2nd Symposium on Advances in Approximate Bayesian Inference.
Chowdhury, Oliveira, and Ramos. 2020.
“Active Learning of Conditional Mean Embeddings via Bayesian Optimisation.” In
Proceedings of the 36th Conference on Uncertainty in Artificial Intelligence (UAI).
Grünewälder, Lever, Baldassarre, et al. 2012.
“Conditional Mean Embeddings as Regressors.” In
Proceedings of the 29th International Coference on International Conference on Machine Learning. ICML’12.
Muandet, Fukumizu, Sriperumbudur, et al. 2017.
“Kernel Mean Embedding of Distributions: A Review and Beyond.” Foundations and Trends® in Machine Learning.
Nishiyama, and Fukumizu. 2016.
“Characteristic Kernels and Infinitely Divisible Distributions.” The Journal of Machine Learning Research.
Pfister, Bühlmann, Schölkopf, et al. 2018.
“Kernel-Based Tests for Joint Independence.” Journal of the Royal Statistical Society: Series B (Statistical Methodology).
Salvi, Lemercier, Liu, et al. 2024.
“Higher Order Kernel Mean Embeddings to Capture Filtrations of Stochastic Processes.” In
Advances in Neural Information Processing Systems. NIPS ’21.
Smola, Gretton, Song, et al. 2007.
“A Hilbert Space Embedding for Distributions.” In
Algorithmic Learning Theory. Lecture Notes in Computer Science 4754.
Song, Huang, Smola, et al. 2009.
“Hilbert Space Embeddings of Conditional Distributions with Applications to Dynamical Systems.” In
Proceedings of the 26th Annual International Conference on Machine Learning. ICML ’09.
Sriperumbudur, B. K., Gretton, Fukumizu, et al. 2008.
“Injective Hilbert Space Embeddings of Probability Measures.” In
Proceedings of the 21st Annual Conference on Learning Theory (COLT 2008).
Sriperumbudur, Bharath K., Gretton, Fukumizu, et al. 2010.
“Hilbert Space Embeddings and Metrics on Probability Measures.” Journal of Machine Learning Research.