Abstract
Selecting optimal number of components in Gaussian Mixture Model (GMM) is of interest to many researchers in the last few decades. Most current approaches are based on information criterion, which was introduced by Akaike (1974) and modified by many other researchers. The standard approach uses the EM algorithm, which fits model parameters to training data and determines log-likelihood functions for increasing number of components. Penalized forms of log-like- lihood function are then used for selecting number of components. Just searching for new or modified forms of penalty function is subject of permanent effort how to improve or do robust these methods for various type of distributed data. Our new technique for selection of optimal number of GMM components is based on Multiple Random Subsampling of training data with Initialization of the EM algorithm (MuRSI). Results of many performed experiments demonstrate the advantages of this method.
This paper was supported by the project no. P103/12/G084 of the Grant Agency of the Czech Republic.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Akaike, H.: A new look at the statistical model identification. IEEE Transactions on Automatic Control AC–19, 716–723 (1974)
Bulteel, K., Wilderjans, T.F., Tuerlinckx, F., Ceulemans, E.: CHull as an alternative to AIC and BIC in the context of mixtures of factor analyzers. Behav. Res. 45, 782–791 (2013)
Chen, P., Wu, T.J., Yang, J.: A comparative study of model selection criteria for the number of signals. IET Radar Sonar Navig. 2(3), 180–188 (2008)
Huang, T., Peng, H., Zhang, K.: Model Selection for Gaussian Mixture Models. Cornell University Library (2013). http://arxiv.org/abs/1301.3558v1
McLachlan, G., Peel, D.: Finite Mixture Models. John Wiley & Sons, New York (2000)
Oliver, Ch., Jouzel, F., El Matouat, A.: Choice of the number of component clusters in mixture models by information criteria. In: Vision Interface 1999, Trois-Riveres, Canada (1999)
Schwarz, G.: Estimating the dimension of a model. Ann. Stat. 6, 461–464 (1978)
Shibata, R.: Selection of the order of an autoregressive model by Akaik’s information criterion. Biometrika 52(3), 333–345 (1976)
Vrieze, S.I.: Model selection and psychological theory: A discussion of the differences between the Akaike Information Criterion (AIC) and the Bayesian Information Criterion (BIC). Psychological Methods 17(2), 228–243 (2012)
Xie, C., Chang, J., Liu, Y.: Estimating the Number of Components in Gaussian Mixture Models Adaptively. Journal of Information & Computational Science 10(14), 4453–4460 (2013)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Psutka, J.V. (2015). Gaussian Mixture Model Selection Using Multiple Random Subsampling with Initialization. In: Azzopardi, G., Petkov, N. (eds) Computer Analysis of Images and Patterns. CAIP 2015. Lecture Notes in Computer Science(), vol 9256. Springer, Cham. https://doi.org/10.1007/978-3-319-23192-1_57
Download citation
DOI: https://doi.org/10.1007/978-3-319-23192-1_57
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-23191-4
Online ISBN: 978-3-319-23192-1
eBook Packages: Computer ScienceComputer Science (R0)