a.k.a. state space model calibration, recursive identification. Sometimes indistinguishable from online estimation.

State filters are cool for estimating time-varying hidden states given known fixed system parameters. How about learning those parameters of the model generating your states? Classic ways that you can do this in dynamical systems include basic linear system identification, and general system identification. But can you identify the fixed parameters (not just hidden states) with a state filter?

Yes.

According to (Lindström et al. 2012), here are some landmark papers:

Augmenting the unobserved state vector is a well known technique, used in the system identification community for decades, see e.g. Ljung (L. Ljung 1979; Lindström et al. 2008; Söderström and Stoica 1988). Similar ideas, using Sequential Monte Carlos methods, were suggested by (Kitagawa 1998; Liu and West 2001). Combined state and parameter estimation is also the standard technique for data assimilation in high-dimensional systems, see Moradkhani et al. (Geir Evensen 2009; G. Evensen 2009; Moradkhani et al. 2005)

However, introducing random walk dynamics to the parameters with fixed variance leads to a new dynamical stochastic system with properties that may be different from the properties of the original system. That implies that the variance of the random walk should be decreased, when the method is used for offline parameter estimation, cf. (Hürzeler and Künsch 2001).

## Classic recursive estimation

🏗

## Iterated filtering

Related: indirect inference. Precise relation will have to wait, since I currently do not care enough about indirect inference.

## Questions

Ionides and King dominate my citations, at least for the frequentist stuff. Surely other people use this method too? But what are the keywords? This research is suspiciously concentrated in U Michigan, but the idea is not so esoteric. I think I am caught in a citation bubble.

Update: the oceanographers, e.g. (Geir Evensen 2003), seem to do this with Bayes a lot.

a lot of the variational filtering literature turns out to be about attempting this with, effectively, neural nets.

can I estimate regularisation this way, despite the lack of probabilistic interpretation? (leveraging Bayesian-prior parameter relations)

How does this work with non-Markov systems? Do we need to bother, or can we just do the Hamiltonian trick and augment the state vector? Can we talk about mixing, or correlation decay? Should I then shoot for the new-wave mixing approaches of Kuznetsov and Mohri etc?

### Basic Construction

There are a few variations. We start with the basic continuous time state space model.

Here we have an unobserved Markov state process \(x(t)\) on \(\mathcal{X}\) and an observation process \(y(t)\) on \(\mathcal{Y}\). For now they will be assumed to be finite dimensional vectors over \(\mathbb{R}.\) They will additionally depend upon a vector of parameters \(\theta\) We observe the process at discrete times \(t(1:T)=(t_1, t_2,\dots, t_T),\) and we write the observations \(y(1:T)=(y(t_1), y(t_2),\dots, y(1_T)).\)

We presume our processes are completely specified by the following conditional densities (which might not have closed-form expression)

The transition density

\[f(x(t_i)|x(t_{i-1}), \theta)\]

The observation density…

TBC.

## Implementations

pomp does state filtering inference in R.

For some example of doing this in Stan see Sinhrks’ stan-statespace.

## Incoming

Recently enjoyed:
Sahani Pathiraja’s state filter does something cool,
in attempting to identify
process *model* noise — a conditional nonparametric density of process errors,
that may be used to come up with some neat process models.
I’m not convinced about her use of
kernel density estimation,
since these scale badly precisely when you need them most,
in high dimension; but any nonparametric density estimator would,
I assume, work, and that would be awesome.

## References

*Journal of the Royal Statistical Society: Series B (Statistical Methodology)*72 (3): 269–342.

*arXiv:1511.07367 [Stat]*, November.

*Proceedings of the National Academy of Sciences*111 (52): 18507–12.

*arXiv:1707.01069 [Cs, Stat]*, July.

*International Conference on Machine Learning*, 544–52.

*The Annals of Applied Statistics*3 (1): 319–48.

*Proceedings of the National Academy of Sciences*113 (15): 3932–37.

*Advances in Neural Information Processing Systems 28*, edited by C. Cortes, N. D. Lawrence, D. D. Lee, M. Sugiyama, and R. Garnett, 2980–88. Curran Associates, Inc.

*Journal of the Royal Statistical Society: Series B (Statistical Methodology)*68 (3): 411–36.

*Statistics and Computing*22 (5): 1009–20.

*Sequential Monte Carlo Methods in Practice*. New York, NY: Springer New York.

*arXiv:1304.5768 [Stat]*, April.

*Bayesian Analysis*11 (2): 325–52.

*Time Series Analysis by State Space Methods*. 2nd ed. Oxford Statistical Science Series 38. Oxford: Oxford University Press.

*IEEE Control Systems*29 (3): 83–104.

*Ocean Dynamics*53 (4): 343–67.

*Data Assimilation - The Ensemble Kalman Filter*. Berlin; Heidelberg: Springer.

*Annual Review of Statistics and Its Application*5 (1): 421–49.

*Journal of The Royal Society Interface*7 (43): 271–83.

*arXiv:1411.5172 [Cs, Stat]*, November.

*Sequential Monte Carlo Methods in Practice*, 159–75. Statistics for Engineering and Information Science. Springer, New York, NY.

*PMLR*, 1607–16.

*Proceedings of the National Academy of Sciences*103 (49): 18438–43.

*The Annals of Statistics*39 (3): 1776–1802.

*Proceedings of the National Academy of Sciences*112 (3): 719–24.

*IFAC Proceedings Volumes*, 15th IFAC Symposium on System Identification, 42 (10): 774–85.

*Statistical Science*30 (3): 328–51.

*Journal of the American Statistical Association*, 1203–15.

*arXiv Preprint arXiv:1511.05121*.

*Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence*, 2101–9.

*arXiv Preprint arXiv:1705.10306*.

*Ecology Letters*10 (7): 551.

*Journal of the American Statistical Association*105 (492): 1617–25.

*IFAC-PapersOnLine (System Identification, Volume 16)*, 45:1785–90. 16th IFAC Symposium on System Identification. IFAC & Elsevier Ltd.

*Computational Statistics & Data Analysis*52 (6): 2877–91.

*Sequential Monte Carlo Methods in Practice*, 197–223. Statistics for Engineering and Information Science. Springer, New York, NY.

*IEEE Transactions on Automatic Control*24 (1): 36–50.

*Stochastic Approximation and Optimization of Random Systems*. Vol. 17. Birkhäuser.

*Theory and Practice of Recursive Identification*. The MIT Press Series in Signal Processing, Optimization, and Control 4. Cambridge, Mass: MIT Press.

*arXiv Preprint arXiv:1705.09279*.

*Advances in Water Resources*28 (2): 135–47.

*arXiv Preprint arXiv:1705.11140*.

*arXiv:1703.00381 [Cs, Stat]*, March.

*Automatica*, Trends in System Identification, 31 (12): 1691–1724.

*System Identification*. Upper Saddle River, NJ, USA: Prentice-Hall, Inc.

*arXiv:1705.08209 [Cs]*, May.

*Monthly Weather Review*131 (7): 1485–90.

*Neural Networks*1 (4): 339–56.

## No comments yet. Why not leave one?