The Matrix-Gaussian distribution

August 19, 2022 — February 15, 2023

algebra
geometry
high d
linear algebra
measure
probability
signal processing
spheres
statistics
Figure 1

Gupta and Nagar ():

The random matrix X(p×n) is said to have a matrix variate normal distribution with mean matrix M(p×n) and covariance matrix ΣΨ where Σ(p×p)>0 and Ψ(n×n)>0, if vec(X)Npn(vec(M),ΣΨ)

We shall use the notation XNp,n(M,ΣΨ).

They prove the following theorem:

If XNp,n(M,ΣΨ), then the p.d.f. of X is given by (2π)12npdet(Σ)12ndet(Ψ)12petr{12Σ1(XM)Ψ1(XM)}XRp×n,MRp×n

Is this the same matrix normal as discussed in scipy.stats.matrix_normal? If so

The probability density function for matrix_normal is f(X)=(2π)mn2|U|n2|V|m2exp(12Tr[U1(XM)V1(XM)T]) where M is the mean, U the among-row covariance matrix, V the among-column covariance matrix. The allow_singular behaviour of the multivariate_normal distribution is not currently supported. Covariance matrices must be full rank. The matrix_normal distribution is closely related to the multivariate_normal distribution. Specifically, Vec(X) (the vector formed by concatenating the columns of X) has a multivariate normal distribution with mean Vec(M) and covariance VU (where is the Kronecker product). Sampling and pdf evaluation are O(m3+n3+m2n+mn2) for the matrix normal, but O(m3n3) for the equivalent multivariate normal, making this equivalent form algorithmically inefficient.

Looks right. For an actual introduction, the section in The Book of Statistical Proof proves some useful theorems in a consistent notation.

1 References

Gupta, and Nagar. 1999. Matrix Variate Distributions. Chapman & Hall/CRC Monographs and Surveys in Pure and Applied Mathematics 104.