I have a nonlinear transformation of a random process. What is its distribution?

Related: What is the gradient of the transform? That is the topic of the reparameterization trick and other MC grad estimators. This is very commonly seen in the context of transforms of Gaussians when it produces results like the delta method and extended Kalman filtering.

## Taylor expansion

Not complicated but subtle (Gustafsson and Hendeby 2012).

Consider a general nonlinear differentiable transformation \(g\) and its second order Taylor expansion. Consider the mapping \(g:\mathbb{R}^{n_{x}}\to\mathbb{R}^{n_{z}}\) applied to a variable \(x,\) defining \(z:=g(x).\) Let \(\mathrm{E}(x)=\mu_{x}\) and \(\operatorname{Var}(x)=P_{x}.\) The Hessian of the \(i^{\text {th }}\) component of \(g\) is denoted \(g_{i}^{\prime \prime}.\) \([x_i]_i\) is a vector where the \(i\)th element is \(x_i\). We will approximate \(z\) using the Taylor expansion, \[z=g\left(\mu_{x}\right)+g^{\prime}\left(\mu_{x}\right)\left(x-\mu_{x}\right)+\left[\frac{1}{2}\left(x-\mu_{x}\right)^{T} g_{i}^{\prime \prime}\left(\mu_{x}\right)\left(x-\mu_{x}\right)\right]_{i}.\] Leaving aside questions of when this is convergent for now. Then the first moment of \(z\) is given by \[ \mu_{z}=g\left(\mu_{x}\right)+\frac{1}{2}\left[\operatorname{tr}\left(g_{i}^{\prime \prime}\left(\mu_{x}\right) P_{x}\right)\right]_{i} \] Further, let \(x \sim \mathcal{N}\left(\mu_{x}, P_{x}\right)\), then the second moment of \(z\) is given by \[ P_{z}=g^{\prime}\left(\mu_{x}\right) P_{x}\left(g^{\prime}\left(\mu_{x}\right)\right)^{T}+\frac{1}{2}\left[\operatorname{tr}\left(g_{i}^{\prime \prime}\left(\mu_{x}\right) P_{x} g_{j}^{\prime \prime}\left(\mu_{x}\right) P_{x}\right)\right]_{i j} \] with \(i, j=1, \ldots, n_{z}.\)

This is commonly seen in the context of transforms of Gaussians.

## Unscented transforms

Typically seen for Gaussian RVs. Ebeigbe et al. (2021) claims to have devised a method for more general RVs.

## Stein’s lemma

As seen in Stein’s method. Gives us the special case of certain exponential RVS (typically Gaussian) under certain matched transforms. Long story.

## Stochastic Itô-Taylor expansion

Taylor expansions for stochastic processes.
See stochastic taylor expansion.
**tl;dr**: Usually more trouble than it is worth.

## Learnable transforms

## References

*Handbook of Financial Econometrics: Tools and Techniques*, 1–66. Elsevier.

*Malaysian Journal of Fundamental and Applied Sciences*13 (3).

*Journal of Economic Dynamics and Control*25 (6-7): 979–99.

*arXiv:2006.13429 [Cs, Math]*, June.

*ArXiv*, April, arXiv:2104.01958v1.

*2008 IEEE International Conference on Acoustics, Speech and Signal Processing*, 3617–20.

*IEEE Transactions on Signal Processing*60 (2): 545–55.

*Lévy Processes: Theory and Applications*, edited by Ole E. Barndorff-Nielsen, Sidney I. Resnick, and Thomas Mikosch, 139–68. Boston, MA: Birkhäuser.

*Mathematische Nachrichten*151 (1): 33–50.

*Stochastic Analysis and Applications*10 (4): 431–41.

*Numerical Solution of Stochastic Differential Equations*, edited by Peter E. Kloeden and Eckhard Platen, 161–226. Applications of Mathematics. Berlin, Heidelberg: Springer.

*Numerical Solution of Stochastic Differential Equations*. Berlin, Heidelberg: Springer Berlin Heidelberg.

*arXiv:1910.13398 [Cs, Stat]*, October.

*Heliyon*5 (2): e01136.

*arXiv:0906.5581 [Math, q-Fin]*, October.

*Stochastic Analysis and Applications*22 (6): 1553–76.

*Journal of Advances in Information Fusion*11 (1): 47–70.

*Journal of Economic Dynamics and Control*28 (4): 755–75.

*Introduction to Variance Estimation*. 2nd ed. Statistics for Social and Behavioral Sciences. New York: Springer.

## No comments yet. Why not leave one?