Ensemble Kalman updates are empirical Matheron updates
June 22, 2022 — February 6, 2025
Assumed audience:
Machine learning researchers and geospatial statisticians, together in peace at last
A note about a fact that I use a lot, which is that the Ensemble Kalman Filter can be formulated as, and IMO more conveniently explained as, an empirical version of the Matheron update, which is a handy tool from Gaussian process regression.
If you prefer your maths in PDF form, check out the arXiv report that this post is translated from (MacKinlay 2025). The fact that this was translated from a semi-formal paper might also warn you that the language here will be stilted and jargony. I may fix that if I have time some day.
This is one of those facts that I am amazed does not seem to be widely known. I would be surprised if I were the first person to notice it, but on the other hand, I do not have another citation that lays out the connection directly, and it is very useful to know. It may be in one of the references below; I have not read all of them, but rather given up after looking at too many candidates after I came up blank. Please let me know in the comments if you complete that quest.
1 Introduction
The Ensemble Kalman Filter (EnKF) (Evensen 2003, 2009a) is a cornerstone in data assimilation for large-scale dynamical systems due to its computational efficiency and scalability. The EnKF approximates the state estimation problem by evolving an ensemble of state vectors through the model dynamics and updating them using observational data.
Separately, the Matheron update provides a sample-based method for conditioning Gaussian random variables on observations(Doucet 2010; Wilson et al. 2020, 2021). We transform prior samples into posterior samples, which is sometimes much easier than calculating posterior densities. This approach is well-established in geostatistics and spatial statistics, but the connection to ensemble methods in data assimilation seems not to be well-known.
In this work, we establish that the ensemble update step in the EnKF is equivalent to an empirical Matheron update by putting them on a common probabilistic footing. By explicitly representing the ensemble mean and covariance using empirical approximations, we demonstrate this equivalence. Recognising this connection provides an alternative probabilistic foundation for the EnKF and suggests potential improvements in ensemble data assimilation techniques by leveraging the properties of the Matheron update. Conversely, the analytic Matheron updates in the literature could benefit from the computational optimisations of the data assimilation community.
1.1 Notation
We write random variates sans serif,
2 Matheron Update
The Matheron update is a technique for sampling from the conditional distribution of a Gaussian random variable given observations, without explicitly computing the posterior covariance (Doucet 2010; Wilson et al. 2020, 2021) but instead updating prior samples. Spoiler: the EnKF update is the same thing for a particular kind of Gaussian model where the posterior is generated by a simulator.
Suppose we have a jointly Gaussian vector
Then the conditional distribution
and
The Matheron update (Doucet 2010) says that drawing a sample from
3 Kalman Filter
We begin by recalling that in state filtering the goal is to update our estimate of the system state when a new observation becomes available. In a general filtering problem, the objective is to form the posterior distribution
We assume a known observation operator
The analysis step—in state filtering parlance—is the update at time
Suppose that the state and observation noise are independent, and all variates are defined over a finite-dimensional real vector space, with
Then the joint distribution of the prior state and observation is Gaussian, and equation Equation 5 implies that it is
When an observation
The Kalman gain is defined by
4 Ensemble Kalman Filter
In high-dimensional or nonlinear settings, directly computing these posterior updates is often intractable. The Ensemble Kalman Filter (EnKF) addresses this by representing the belief about the state empirically via an ensemble of
Since Gaussian measures are completely specified by their first two moments, we construct empirical measures that match the target in these moments. For convenience, we introduce the following notation for the matrix mean and deviations:
The ensemble mean (see Equation 8) and covariance (see Equation 9) are computed from the empirical measure:
Abusing notation, we associate a Gaussian distribution with the empirical measure:
We overload the observation operator to apply to ensemble matrices, writing
Here the regularization term
The Kalman gain in the ensemble setting is constructed by plugging the empirical estimates from Equation 10 into Equation 6. This yields
5 Empirical Matheron Update is Equivalent to Ensemble Kalman Update
Under the substitution
6 Computational Complexity
As is well understood in the EnKF literature, one of the main advantages of the EnKF is that the analysis update can be performed using only empirical ensemble statistics, thereby avoiding the direct computation and inversion of full, high-dimensional covariance matrices. The naive Kalman update must operate on these full covariance matrices, which can be computationally prohibitive.
In practice, we calculate the gain from Equation 11 as follows:
Thus, the overall computational cost of the ensemble update is
7 Implications
Understanding the EnKF as an empirical Matheron update opens up several possibilities. Both the Matheron update and Ensemble Kalman methods have been extensively studied in their respective fields. In recent years, numerous developments in Bayesian machine learning have advanced both approaches; see the many developments in Ensemble Kalman methods in the Ensemble Kalman notebook and especially the Neural networks meet Kalman filters notebook.
If we interpret the pathwise Matheron update in the context of the empirical measures used in the EnKF, it suggests the potential to import developments from one field to the other. The vast literature on the EnKF indicates that many ideas developed there might be applicable to Gaussian Process regression via the Matheron update. For instance, the clear probabilistic interpretation of the ensemble update could inspire improved regularization and localization strategies. One example is the Local Ensemble Transform Kalman Filter (Bocquet, Farchi, and Malartic n.d.), which provides state-of-the-art performance in high-dimensional filtering and system identification settings and could be adapted to the Matheron update (homework exercise).
Where to go from here? I don’t know; but cite me if this makes your life easier.
8 References
Footnotes
This measure notation from probability theory is unpopular in both machine learning and data assimilation contexts, but is succinct here.↩︎