Mixture models for density estimation

March 29, 2016 — April 24, 2020

Bayes
classification
clustering
compsci
convolution
density
information
linear algebra
nonparametric
probability
sparser than thou
statistics
Figure 1

A method of semi-parametric density estimation.

We consider random variables \(\mathsf{x}\) with density made out of a mixture of other component densities.

\[ f(x) \sim \sum_{i = 1}^M w_i f(x;\theta_i). \] Here \(\sum_i w=1\) and \(f(x;\theta_i)\) are the mixture component densities, which are themselves specified by some parameter vector \(\theta_i.\) These components do not need to be in the same family, so let us assume that the \(\theta_i\) can encode some information about which family too. Equivalently, we consider the variables defined by randomly choosing from an array of mutually independent component random variables \(\mathsf{x}_i\sim f(\cdot; \theta_i)\).

\[ \mathsf{x} = \begin{cases} \vdots \\ \mathsf{x}_i: \text{ with prob. } w_i\\ \vdots \\ \end{cases} \] We might imagine there is some latent variable \(\mathsf{w}\sim \mathrm{Multinomial}(w_1,\dots,w_M)\) which we use to choose the component. This seems to be the formalism we use more commonly in model averaging.

There are many connections. Inferring mixture models is conceptually similar to clustering, and indeed many papers note the connection. Clustering via nearest-centroids is more or less mixture modelling where the mixture components all have the same shape. The classic inference method is expectation maximisation, which is of independent interest. A degenerate case, convolution kernel density estimation leads to particularly simple results. Model selection and convergence theory is a gigantic mess which I will not find time to disentangle.

Figure 2: pyxelate uses mixture models to create pixel art colour palettes

1 Moments of a mixture

Before we go on to consider all the knotty convoluted estimation and approximation theory, let us think about the basic probabilistic properties of mixtures.

\[ \begin{aligned} E(\mathsf{x}) = {} & E[E(\mathsf{x}\mid \mathsf{w})] \\ = {} & E\left.\begin{cases} \vdots \\ E[\mathsf{x}_i] & \text{with prob. }w_i \\ \vdots \end{cases}\right\} \\ = {} & \sum_{i=1}^M w_i \mu_i\\ =: {} & \bar{\mu}. \end{aligned}\]

\[ \begin{aligned} \var(\mathsf{x}) = {} & E(\var(\mathsf{x}\mid \mathsf{w})) + \var(E(\mathsf{x} \mid \mathsf{w})) \\ = {} & E\left.\begin{cases} \vdots \\ \var[\mathsf{x}_i] & \text{with prob. }w_i \\ \vdots \end{cases}\right\} \\ & {} + \var\left.\begin{cases} \vdots \\ E[\mathsf{x}_i] & \text{with prob. }w_i \\ \vdots \end{cases} \right\} \\ = {} & \sum_{i=1}^M w_i \var(\mathsf{x}_i) + \sum_{i=1}^M w_i(\mu_i-\bar\mu)(\mu_i-\bar\mu)^T, \end{aligned}\] where \(\mu_i:=E(\mathsf{x}_i)\) and \(\bar{\mu}\) is the weighted mean over all \(\mu_i.\)

Nice. This simplicity is why you would work with these creatures, out of all the crazy ways we have of combining random variables into new random variables (e.g. reparameterization), mixtures are the ones that retain

2 Mixture zoo

The following categories are not mutually exclusive; In fact I’m mentioning them all to work out what exactly the intersections are.

(Battey and Liu 2013; van de Geer 1996; Assaf J. Zeevi and Meir 1997) discuss some useful results common to various convex combination estimators.

(DasGupta 2008) ch 33-24 is a high-speed no-filler all-killer summary of various convergence results and mixture types, including a connection to Donoho-Jin “Higher criticism” and nonparametric deconvolution and multiple testing.

2.1 “Classic Mixtures”

Finite location-scale mixtures, usually of Gaussians.

Your data are vectors \(x_j\in \mathbb{R}^d\).

Your density looks like this:

\[ f(x_j) = w_0 + \sum_{i=1}^{i=m}w_i\phi((x_j-\mu_j)^T\sigma_i^{-1}(x_j-\mu_j))/|\sigma| \]

Traditionally, \(\phi\) is given, and given as the normal density, but any “nice” elliptic density will do.

Also traditionally, \(m\) is given by magic. Fitting the parameters of this model, then, involves choosing \((\mathbf{w},\boldsymbol{\mu},\mathbf{\sigma}),\, \mathbf{w}\in\mathbb{R}^{m+1}, \boldsymbol{\mu}\in\mathbb{R}^{m\times d}, \boldsymbol{\sigma}\in\mathbb{R}^{d\times d}.\)

Why would you bother? We know that this class is dense in the space of all densities under the total variation metric (Cheney and Light 2009), which is a rationale if not a guarantee of its usefulness. For the Gaussian one can appeal to, e.g. (Battey and Liu 2013; Assaf J. Zeevi and Meir 1997) to argue that we can get “close” to any density in large classes by choosing the Gaussian for big \(m\) and introducing some strategic assumptions. Moreover, we know that it’s computationally tractable.

2.1.1 Radial basis functions

A finite mixtures subtype, from the function approximation literature. When your component density has an assumed form

\[ f(x_j) = w_0 + \sum_{i=1}^{i=m}w_i\phi(h_j\|x_j-\mu_j\|)/h_j \]

Here \(h\) is a scale parameter for the density function. This corresponds to a spherical, rather than elliptic, approximating function, so has fewer parameters per component in higher dimensions than one.

2.1.2 Kernel density estimators

If you have as many mixture components as data points, you have kernel density estimate. This is a semiparametric limiting case of the radial basis function model. To keep the number of parameters manageable you usually assume that the mixture components themselves all share the same scale parameters, although there are some extensions to that around which I might recall if I have need.

2.2 Continuous mixtures

Noticing that a classic location mixture is a convolution between a continuous density and an atomic density, the question arises whether you can convolve two more general densities. Yes, you can. Estimate a nonparametric mixing density. Now you have a nonparametric estimation problem, something like, estimate \(dP(\mu,\sigma)\). I should mention one interesting sub-case:

2.2.1 Normal variance mixtures

Mixtures of normals where you only vary scale parameters. I clearly don’t know enough about this to write the entry; this is just a note to myself. 🏗 z-distributions and generalized hyperbolic distributions are the keywords. They include many other distributions as special cases.

2.3 Bayesian Dirichlet mixtures

🏗; Something about using a Dirichlet process for the… weights of the mixture components? Giving you a posterior distribution over countably finite mixture parameters? something like that?

2.4 Non-affine mixtures

Do the mixture components have to have location/scale parameterisations? Most models mentioned so far have assumed that, but more general models are feasible. However, estimating the scale parameters is already a PITA; what can we say about more general shape parameters? I suspect this is one way you could approach tail cutoffs in extreme value theory for example.

3 In Bayesian variational inference

Gaussian mixture models are the approximation densities in “classic” flavoured Bayesian variational inference, as opposed to fancy modern neural variational inference. Here you approximate a posterior with a mixture distribution; The estimation method looks reasonably similar to the frequentist case.

4 Estimation/selection methods

4.1 (Local) maximum likelihood

A classic method; There are some subtleties since the global maximum can be badly behaved; you have to mess around with local roots of the likelihood equation and thereby lose some of the lovely asymptotics of MLE methods in exponential families.

However, I am not sure which properties you lose. (Geoffrey J. McLachlan and Rathnayake 2014), for example, makes the assertion that the AIC conditions (whatever they are) don’t hold, but the BIC ones do.

4.2 Method of moments

Particularly popular in recent times for mixtures. 🏗

4.3 Minimum distance

Minimise the distance between the empirical PDF and the estimated PDF in some metric. For reasons I have not yet digested, one is probably best to do this in the Hellinger metric if one wishes convenient convergence, (Beran 1977), although kernel density estimates tend to prefer \(L_2\), as with e.g, regression smoothing problems. How on earth you numerically minimise Hellinger distance from data is something I won’t think about for now, although I admit to being curious.

4.4 Regression smoothing formulation

Not quite mixture models; you have a smoothness penalty on the quadratic component, a log-quadratic regression spline, then the results are “nearly” Gaussian mixture models. See (Eilers and Marx 1996).

5 Convergence and model selection

I would like a convenient text summarising the convergence results for mixture models under varying assumptions. There is a giant heap of research on convergence under various assumptions and metrics which needs summarising. That is beyond my capacity at the moment.

Dustin Mixon explains his paper and a few others in the field, in The Voronoi Means Conjecture

As I mentioned in a previous blog post, I have a recent paper with Soledad Villar and Rachel Ward that analyzes an SDP relaxation of the k-means problem. It turns out that the solution to the relaxation can be interpreted as a denoised version of the original dataset, sending many of the data points very close to what appear to be k-means-optimal centroids. This is similar in spirit to Dasgupta’s random projection, and we use a similar rounding scheme to estimate the optimal k-means-centroids. Using these centroid estimates as Gaussian center estimates, we are able to prove performance bounds of the form \(\mathrm{MSE}\lesssim k^2\sigma^2\) when \(\mathrm{SNR}>rsim k^2\), meaning the performance doesn’t depend on the dimension, but rather the model order.

But does it make sense that the performance should even depend on the model order?

Also van de Geer (1996), leverages the convex combination structure to get bounds relating KL and Hellinger convergence of density estimators (including kernel density estimators)

(Dasgupta 1999) probably counts, and (Mixon, Villar, and Ward 2016) introduces others, but most of these seem not to address the approximation problem. but the clustering problem. Clustering doesn’t fit perfectly with my purpose here; I don’t necessarily care about assigning things correctly to clusters; rather, I want to approximate the overall density well.

Mixon, Villar, and Ward (2016) has for some interesting connections to matrix concentration inequalities:

The study of theoretical guarantees for learning mixtures of Gaussians started with the work of (Dasgupta 1999). His work presented an algorithm based on random projections and showed this algorithm approximates the centres of Gaussians […] in \(R^m\) separated by […] the biggest singular value among all [covariance matrices]. After this work, several generalisations and improvements appeared. […] To date, techniques used for learning mixtures of Gaussians include expectation maximisation (Dasgupta and Schulman 2007), spectral methods (Awasthi and Sheffet 2012; Kumar and Kannan 2010), projections (random and deterministic) (Dasgupta 1999; Moitra and Valiant 2010; Arora and Kannan 2001), and the method of moments (Moitra and Valiant 2010).

Every existing performance guarantee exhibits one of the following forms:

  • the algorithm correctly clusters all points according to Gaussian mixture component, or

  • the algorithm well-approximates the centre of each Gaussian (a la(Dasgupta 1999).)

Results of type (1), which include (Awasthi and Sheffet 2012; Kumar and Kannan 2010; Achlioptas and Mcsherry 2007), require the minimum separation between the Gaussians centres to have a multiplicative factor of polylog N, where N is the number of points. This stems from a requirement that every point be closer to their Gaussian’s centre (in some sense) than the other centres, so that the problem of cluster recovery is well-posed. Note that in the case of spherical Gaussians, the Gaussian components can be truncated to match the stochastic ball model in this regime, where the semidefinite program we present is already known to be tight with high probability (Iguchi et al. 2015; Awasthi et al. 2015). Results of type (2) tend to be specifically tailored to exploit unique properties of the Gaussians, and as such are not easily generalisable to other data models. […] For instance, if \(\mathsf{x} \sim N (\mu, \sigma^2 Im )\), then \(E(\|\mathsf{x} − \mu\|^2) = m\sigma^2\). In high dimensions, since the entries of the Gaussians are independent, concentration of measure implies that most of the points will reside in a thin shell with center μ and radius about \(\sqrt{m\sigma}\). This property allows algorithms to cluster even concentric Gaussians as long as the covariances are sufficiently different. However, algorithms that allow for no separation between the Gaussian centers require a sample complexity which is exponential in k (Moitra and Valiant 2010).

Hmm.

Here’s one lately-popular extension to the finite mixture model: Choosing the number of mixture components adaptively, using some kind of model selection procedure, as per (Priebe and Marchette 2000) with the “sieve” — (Murata, Yoshizawa, and Amari 1994) uses an information criterion.

There are various optimality results, depending on your assumptions on mixing distributions and component distributions and true distributions and the constellation and so on.

AFAICT all assume either fixed component scales, or require a free regularisation parameter selected by cross-validation.

I would like sample complexity results that can allow me to bound the number of components before seeing the data. That would have been handy for a research project of mine, but I am unaware of any, and have since discontinued that project. All that remains is this bestiary of approaches.

5.1 Large sample results for mixtures

I would prefer not to rely on asymptotic distributions, but sometimes they help.

(Geoffrey J. McLachlan and Rathnayake 2014) claims the AIC doesn’t work because ‘certain regularity assumptions’ necessary for it are not satisfied (which?) but that one may try the BIC, as per (A. R. Barron and Cover 1991; Andrew R. Barron et al. 2008). Contrariwise, (Massart 2007) gives explicit cases where the AIC does work in penalised density estimation, so… yeah. In any case, it’s certainly true that the derivation and calculation of the GIC is tedious.

(Ghosal and van der Vaart 2001) give MLE risk bounds under various assumptions, which I could maybe use independently.

5.2 Finite sample results for mixtures

In a penalised least-squares framework, see (Bertin, Pennec, and Rivoirard 2011). This generally seems much more reasonable to me, and I’m reading up on it. However, I think that basis selection is going to turn out to be nasty.

Massart (Massart 2007) explains some of the complications and gives example applications of both least-squares and ML model selection procedures for general density model selection. However, his results are restricted to mixtures of orthonormal bases, which, realistically, means histograms with bins of disjoint support, if you want your density to be non-negative. Wondering how much work it is to get overlapping support.

5.3 Sieve method

Argh! so many variants. What I would like for my mixture sieve

(Priebe 1994; Genovese and Wasserman 2000)), Battey

5.4 Akaike Information criterion

Use an Akaike-type information criterion

See (Ando, Konishi, and Imoto 2008; Andrew R. Barron et al. 2008; A. Barron, Rissanen, and Yu 1998; Murata, Yoshizawa, and Amari 1994)

(Konishi and Kitagawa 2008) 6.1. is a compressed summary intro to general regularized basis expansion in a regression setting, i.e. we approximate an arbitrary function. Density approximation is more constrained, since we know that our mixture must integrate to 1. Also we don’t have a separate error term; rather, we assume our components completely summarize the randomness. Usually, although not always, we further require the components be non-negative functions with non-negative weights to give us specifically a convex combination of functions. Anyway, presumably we can extract a similar result from that?

5.5 Quantization and coding theory

The information theoretic cousin. Non-uniform quantisation in communication theory is when you optimally distribute the density of your quantization symbols according to the density of the signal, in order to compress a signal while still reconstructing it as precisely as possible. This connection is most commonly raised in the multidimensional case, when it is “Vector quantization” or VQ to its friends. See, e.g. (Gersho and Gray 2012). This is then a coding theory problem.

From reading the literature it is not immediately apparent how, precisely, vector quantisation is related to mixture density estimation, although there is a family resemblance. In vector quantisation you do something like reduce the signal to a list of Voronoi cells and the coordinates of their centres, then code a signal to the nearest centre; Squinting right makes this look like a mixture problem. (D. D. Lee and Seung 2001, 1999) make this connection precise. Investigate.

Now, how do you choose this optimal code from measurements of the signal? THAT is the statistical question.

See quantization for more.

5.6 Minimum description length/BIC

Related to both the previous, in some way I do not yet understand.

Rissanen’s Minimum Description length, as applied to mixture density estimation? Putatively related to the Bayesian Information Criterion, which is purportedly an MDL measure. (?) That is not immediately obvious to me; I thought BIC was just a way of maximising evidence marginal likelihood.

Andrew Barron and co-workers seem to own the statistical MDL approach to mixture estimation. See, (Andrew R. Barron 1994, 1991; A. R. Barron and Cover 1991; A. Barron, Rissanen, and Yu 1998; A. R. Barron 1993), with literature reviews in Andrew R. Barron et al. (2008). That last one constructs discretized mixture models as “two stage codes”, and achieves prediction risk bounds for finite samples using them.

6 Unsatisfactory thing: scale parameter selection theory

Many really good results take the scale parameter as given.

This is fine, as far as it goes, but the scale parameter selection is the reason why I might bother with this class of model, typically, otherwise this is simply a weird convex deconvolution problem, which is not so interesting. In particular, how to handle the scale parameter selection in model selection?

(Belkin, Rademacher, and Voss 2016; Shimazaki and Shinomoto 2010) make a start.

7 Connection to Mercer kernel methods

Connection with kernel PCA (Schölkopf et al. 1998) and metric multidimensional scaling (Williams 2001) and such are explored in kernel approximation.

8 Incoming

Expectation-Maximization with Less Arbitrariness.

9 References

Achlioptas, and McSherry. 2005. On Spectral Learning of Mixtures of Distributions.” In Learning Theory. Lecture Notes in Computer Science.
Achlioptas, and Mcsherry. 2007. Fast Computation of Low-Rank Matrix Approximations.” J. ACM.
Anandkumar, Hsu, and Kakade. 2012. A Method of Moments for Mixture Models and Hidden Markov Models. In Journal of Machine Learning Research.
Ando, Konishi, and Imoto. 2008. Nonlinear Regression Modeling via Regularized Radial Basis Function Networks.” Journal of Statistical Planning and Inference, Special Issue in Honor of Junjiro Ogawa (1915 - 2000): Design of Experiments, Multivariate Analysis and Statistical Inference,.
Arora, and Kannan. 2001. Learning Mixtures of Arbitrary Gaussians.” In Proceedings of the Thirty-Third Annual ACM Symposium on Theory of Computing. STOC ’01.
Awasthi, Bandeira, Charikar, et al. 2015. Relax, No Need to Round: Integrality of Clustering Formulations.” In Proceedings of the 2015 Conference on Innovations in Theoretical Computer Science. ITCS ’15.
Awasthi, and Sheffet. 2012. Improved Spectral-Norm Bounds for Clustering.” arXiv Preprint arXiv:1206.3204.
Bach. 2014. Breaking the Curse of Dimensionality with Convex Neural Networks.” arXiv:1412.8690 [Cs, Math, Stat].
Barron, Andrew R. 1991. Complexity Regularization with Application to Artificial Neural Networks.” In Nonparametric Functional Estimation and Related Topics. NATO ASI Series 335.
Barron, A.R. 1993. Universal Approximation Bounds for Superpositions of a Sigmoidal Function.” IEEE Transactions on Information Theory.
Barron, Andrew R. 1994. Approximation and Estimation Bounds for Artificial Neural Networks.” Machine Learning.
Barron, A. R., and Cover. 1991. Minimum Complexity Density Estimation.” IEEE Transactions on Information Theory.
Barron, Andrew R., Huang, Li, et al. 2008. MDL, Penalized Likelihood, and Statistical Risk.” In Information Theory Workshop, 2008. ITW’08. IEEE.
Barron, A., Rissanen, and Yu. 1998. The Minimum Description Length Principle in Coding and Modeling.” IEEE Transactions on Information Theory.
Battey, and Linton. 2014. Nonparametric Estimation of Multivariate Elliptic Densities via Finite Mixture Sieves.” Journal of Multivariate Analysis.
Battey, and Liu. 2013. Smooth Projected Density Estimation.” arXiv:1308.3968 [Stat].
———. 2016. Nonparametrically Filtered Parametric Density Estimation.”
Battey, and Sancetta. 2013. Conditional Estimation for Dependent Functional Data.” Journal of Multivariate Analysis.
Belkin, Rademacher, and Voss. 2016. Basis Learning as an Algorithmic Primitive.” In Journal of Machine Learning Research.
Bengio, Roux, Vincent, et al. 2005. Convex Neural Networks.” In Advances in Neural Information Processing Systems.
Beran. 1977. Minimum Hellinger Distance Estimates for Parametric Models.” The Annals of Statistics.
Bertin, Pennec, and Rivoirard. 2011. Adaptive Dantzig Density Estimation.” Annales de l’Institut Henri Poincaré, Probabilités Et Statistiques.
Birgé, and Rozenholc. 2006. How Many Bins Should Be Put in a Regular Histogram.” ESAIM: Probability and Statistics.
Bishop, Chris. 1991. Improving the Generalization Properties of Radial Basis Function Neural Networks.” Neural Computation.
Bishop, Christopher. 1994. Mixture Density Networks.” Microsoft Research.
Boyd, Hastie, Boyd, et al. 2016. Saturating Splines and Feature Selection.” arXiv:1609.06764 [Stat].
Chen. 1995. Optimal Rate of Convergence for Finite Mixture Models.” The Annals of Statistics.
Cheney, and Light. 2009. A Course in Approximation Theory.
Christian Bauckhage. 2015. Lecture Notes on Data Science: Soft k-Means Clustering.”
Cohen. 1950. Estimating the Mean and Variance of Normal Populations from Singly Truncated and Doubly Truncated Samples.” The Annals of Mathematical Statistics.
Cohen, Jr. 1949. On Estimating the Mean and Standard Deviation of Truncated Normal Distributions.” Journal of the American Statistical Association.
Daniely, Linial, and Saks. 2012. Clustering Is Difficult Only When It Does Not Matter.” arXiv:1205.4891 [Cs].
Dasgupta. 1999. Learning Mixtures of Gaussians.” In Foundations of Computer Science, 1999. 40th Annual Symposium on.
DasGupta. 2008. Asymptotic Theory of Statistics and Probability. Springer Texts in Statistics.
Dasgupta, and Schulman. 2007. A Probabilistic Analysis of EM for Mixtures of Separated, Spherical Gaussians.” Journal of Machine Learning Research.
Dehaene. 2016. Expectation Propagation Performs a Smoothed Gradient Descent.” arXiv:1612.05053 [Stat].
Donoho, Johnstone, Kerkyacharian, et al. 1996. Density Estimation by Wavelet Thresholding.” The Annals of Statistics.
Dunson, Pillai, and Park. 2007. Bayesian Density Regression.” Journal of the Royal Statistical Society: Series B (Statistical Methodology).
Eilers, and Marx. 1996. Flexible Smoothing with B-Splines and Penalties.” Statistical Science.
Fan. 1991. On the Optimal Rates of Convergence for Nonparametric Deconvolution Problems.” The Annals of Statistics.
Geman, and Hwang. 1982. Nonparametric Maximum Likelihood Estimation by the Method of Sieves.” The Annals of Statistics.
Genovese, and Wasserman. 2000. Rates of Convergence for the Gaussian Mixture Sieve.” Annals of Statistics.
Gersho, and Gray. 2012. Vector Quantization and Signal Compression.
Ghosal, and van der Vaart. 2001. Entropies and Rates of Convergence for Maximum Likelihood and Bayes Estimation for Mixtures of Normal Densities.” The Annals of Statistics.
Gray. 1984. Vector Quantization.” IEEE ASSP Magazine.
Hald. 1949. Maximum Likelihood Estimation of the Parameters of a Normal Distribution Which Is Truncated at a Known Point.” Scandinavian Actuarial Journal.
Hall. 1987. On Kullback-Leibler Loss and Density Estimation.” The Annals of Statistics.
Heigold, Schlüter, and Ney. 2007. On the Equivalence of Gaussian HMM and Gaussian HMM-Like Hidden Conditional Random Fields.” In Eighth Annual Conference of the International Speech Communication Association.
Hinton, Deng, Yu, et al. 2012. Deep Neural Networks for Acoustic Modeling in Speech Recognition: The Shared Views of Four Research Groups.” IEEE Signal Processing Magazine.
Hsu, and Kakade. 2013. Learning Mixtures of Spherical Gaussians: Moment Methods and Spectral Decompositions.” In Proceedings of the 4th Conference on Innovations in Theoretical Computer Science. ITCS ’13.
Huang, Cheang, and Barron. 2008. Risk of Penalized Least Squares, Greedy Selection and L1 Penalization for Flexible Function Libraries.”
Ibragimov. 2001. Estimation of Analytic Functions.” In Institute of Mathematical Statistics Lecture Notes - Monograph Series.
Iguchi, Mixon, Peterson, et al. 2015. Probably Certifiably Correct k-Means Clustering.” arXiv:1509.07983 [Cs, Math, Stat].
Jawitz. 2004. Moments of Truncated Continuous Univariate Distributions.” Advances in Water Resources.
Keriven, Bourrier, Gribonval, et al. 2016. Sketching for Large-Scale Learning of Mixture Models.” In 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
Konishi, and Kitagawa. 2008. Information Criteria and Statistical Modeling. Springer Series in Statistics.
Kumar, and Kannan. 2010. Clustering with Spectral Norm and the k-Means Algorithm.” In 2010 51st Annual IEEE Symposium on Foundations of Computer Science (FOCS).
Lee, Youngjo., Nelder, and Pawitan. 2006. Generalized linear models with random effects. Monographs on statistics and applied probability 106.
Lee, Gyemin, and Scott. 2012. EM Algorithms for Multivariate Gaussian Mixture Models with Truncated and Censored Data.” Computational Statistics & Data Analysis.
Lee, Daniel D., and Seung. 1999. Learning the Parts of Objects by Non-Negative Matrix Factorization.” Nature.
———. 2001. Algorithms for Non-Negative Matrix Factorization.” In Advances in Neural Information Processing Systems 13.
Leung, and Barron. 2006. Information Theory and Mixing Least-Squares Regressions.” IEEE Transactions on Information Theory.
Li, Jonathan Q., and Barron. 2000. Mixture Density Estimation.” In Advances in Neural Information Processing Systems 12.
Li, Dawei, Xu, and Goodman. 2012. On-Line EM Variants for Multivariate Normal Mixture Model in Background Learning and Moving Foreground Detection.” Journal of Mathematical Imaging and Vision.
Li, Haifeng, Zhang, and Jiang. 2005. The Regularized EM Algorithm.” In Proceedings of the National Conference on Artificial Intelligence.
Marin, Mengersen, and Robert. 2005. Bayesian Modelling and Inference on Mixtures of Distributions.” In Handbook of Statistics.
Massart. 2007. Concentration Inequalities and Model Selection: Ecole d’Eté de Probabilités de Saint-Flour XXXIII - 2003. Lecture Notes in Mathematics 1896.
McLachlan, G. J., and Jones. 1988. Fitting Mixture Models to Grouped and Truncated Data via the EM Algorithm.” Biometrics.
McLachlan, Geoffrey J., and Rathnayake. 2014. On the Number of Components in a Gaussian Mixture Model.” Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery.
Mixon, Villar, and Ward. 2016. Clustering Subgaussian Mixtures by Semidefinite Programming.” arXiv:1602.06612 [Cs, Math, Stat].
Moitra, and Valiant. 2010. Settling the Polynomial Learnability of Mixtures of Gaussians.” In 2010 51st Annual IEEE Symposium on Foundations of Computer Science (FOCS).
Murata, Yoshizawa, and Amari. 1994. Network Information Criterion-Determining the Number of Hidden Units for an Artificial Neural Network Model.” IEEE Transactions on Neural Networks.
Norets. 2010. Approximation of Conditional Densities by Smooth Mixtures of Regressions.” The Annals of Statistics.
Orr, Mark JL. 1996. Introduction to Radial Basis Function Networks.”
Orr, Mark J. L. 1999. “Recent Advances in Radial Basis Function Networks.”
Peng, and Wei. 2007. Approximating K‐means‐type Clustering via Semidefinite Programming.” SIAM Journal on Optimization.
Priebe. 1994. Adaptive Mixtures.” Journal of the American Statistical Association.
Priebe, and Marchette. 2000. Alternating Kernel and Mixture Density Estimates.” Computational Statistics & Data Analysis.
Qian, and Chen. 1994. Signal Representation Using Adaptive Normalized Gaussian Functions.” Signal Processing.
Que, and Belkin. 2016. Back to the Future: Radial Basis Function Networks Revisited.” In Proceedings of the 19th International Conference on Artificial Intelligence and Statistics (AISTATS) 2016.
Rabusseau, and Denis. 2014. Learning Negative Mixture Models by Tensor Decompositions.” arXiv:1403.4224 [Cs].
Raghunathan, Krishnaswamy, and Jain. 2017. Learning Mixture of Gaussians with Streaming Data.” In Advances in Neural Information Processing Systems.
Rakhlin, Panchenko, and Mukherjee. 2005. Risk Bounds for Mixture Density Estimation.” ESAIM: Probability and Statistics.
Redner, and Walker. 1984. Mixture Densities, Maximum Likelihood and the EM Algorithm.” SIAM Review.
Rissanen. 1984. Universal Coding, Information, Prediction, and Estimation.” IEEE Transactions on Information Theory.
Roeder, and Wasserman. 1997. Practical Bayesian Density Estimation Using Mixtures of Normals.” Journal of the American Statistical Association.
Ruan, Yuan, and Zou. 2011. Regularized Parameter Estimation in High-Dimensional Gaussian Mixture Models.” Neural Computation.
Schölkopf, Knirsch, Smola, et al. 1998. Fast Approximation of Support Vector Kernel Expansions, and an Interpretation of Clustering as Approximation in Feature Spaces.” In Mustererkennung 1998. Informatik Aktuell.
Schölkopf, Smola, and Müller. 1997. Kernel Principal Component Analysis.” In Artificial Neural Networks — ICANN’97. Lecture Notes in Computer Science.
Sha, and Saul. 2006. Large Margin Hidden Markov Models for Automatic Speech Recognition.” In Advances in Neural Information Processing Systems.
Shimazaki, and Shinomoto. 2010. Kernel Bandwidth Optimization in Spike Rate Estimation.” Journal of Computational Neuroscience.
Singh, and Gordon. 2008. A Unified View of Matrix Factorization Models.” In Machine Learning and Knowledge Discovery in Databases.
Tipping. 2000. The Relevance Vector Machine.” In Advances in Neural Information Processing Systems.
———. 2001. Sparse Bayesian Learning and the Relevance Vector Machine.” The Journal of Machine Learning Research.
Tipping, and Nh. 2001. Sparse Kernel Principal Component Analysis.” In Advances in Neural Information Processing Systems 13.
Toulis, and Airoldi. 2015. Scalable Estimation Strategies Based on Stochastic Approximations: Classical Results and New Insights.” Statistics and Computing.
Tschannen, and Bölcskei. 2016. Noisy Subspace Clustering via Matching Pursuits.” arXiv:1612.03450 [Cs, Math, Stat].
van de Geer. 1996. Rates of Convergence for the Maximum Likelihood Estimator in Mixture Models.” Journal of Nonparametric Statistics.
———. 1997. Asymptotic Normality in Mixture Models.” ESAIM: Probability and Statistics.
———. 2003. Asymptotic Theory for Maximum Likelihood in Nonparametric Mixture Models.” Computational Statistics & Data Analysis, Recent Developments in Mixture Model,.
van der Vaart, and Wellner. 1996. Weak Convergence and Empirical Processes: With Applications to Statistics.
Vempala, and Wang. 2004. A Spectral Algorithm for Learning Mixture Models.” Journal of Computer and System Sciences.
Weidmann, and Vetterli. 2012. Rate Distortion Behavior of Sparse Sources.” IEEE Transactions on Information Theory.
Williams. 2001. On a Connection Between Kernel PCA and Metric Multidimensional Scaling.” In Advances in Neural Information Processing Systems 13.
Wong, and Shen. 1995. Probability Inequalities for Likelihood Ratios and Convergence Rates of Sieve MLES.” The Annals of Statistics.
Wu, Ying, and Zhou. 2007. Multi-Kernel Regularized Classifiers.” Journal of Complexity.
Xu, Lei, and Jordan. 1996. On Convergence Properties of the EM Algorithm for Gaussian Mixtures.” Neural Computation.
Xu, Jian-Wu, Paiva, Park, et al. 2008. A Reproducing Kernel Hilbert Space Framework for Information-Theoretic Learning.” IEEE Transactions on Signal Processing.
Zeevi, Assaf J., and Meir. 1997. Density Estimation Through Convex Combinations of Densities: Approximation and Estimation Bounds.” Neural Networks: The Official Journal of the International Neural Network Society.
Zeevi, A. J., Meir, and Maiorov. 1998. Error Bounds for Functional Approximation and Estimation Using Mixtures of Experts.” IEEE Transactions on Information Theory.