Abstract
Statistical re-sampling techniques have been used extensively and successfully in the machine learning approaches for generation of classifier and predictor ensembles. It has been frequently shown that combining so called unstable predictors has a stabilizing effect on and improves the performance of the prediction system generated in this way. In this paper we use the re-sampling techniques in the context of Principal Component Analysis (PCA). We show that the proposed PCA ensembles exhibit a much more robust behaviour in the presence of outliers which can seriously affect the performance of an individual PCA algorithm. The performance and characteristics of the proposed approaches are illustrated on a number of experimental studies where an individual PCA is compared to the introduced PCA ensemble.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Pearson, K.: On lines and planes of closest fit to systems of points in space. Philosophical Magazine 2, 559–572 (1901)
Hotelling, H.: Analysis of a complex of statistical variables into principal components. Journal of Educational Psychology 24, 417–441, 498–520 (1933)
Cook, R.D.: Detection of influential observations in linear regression. Technometrics 19, 15–18 (1977)
Dixon, W.J.: Analysis of extreme values. Ann. Math. Stat. 21, 488–506 (1950)
Oja, E.: Neural networks, principal components and subspaces. International Journal of Neural Systems 1(1), 61–68 (1989)
Sanger, D.: Contribution analysis: A technique for assigning responsibilities to hidden units in connectionist networks. Connection Science 1, 115–138 (1989)
Breiman, L.: Bagging predictors. Machine Learning 24, 123–140 (1996)
Schapire, R.E., Freud, Y., Bartlett, P., Lee, W.S.: Boosting the margin: a new explanation for the effectiveness of voting methods. The Annals of Statistics 26(5), 1651–1686 (1998)
Gabrys, B.: Combining neuro-fuzzy classifiers for improved generalisation and reliability. In: Proceedings the Int. Joint Conference on Neural Networks (IJCNN 2002) a part of the WCCI 2002 Congress, Honolulu, USA, pp. 2410–2415 (2002)
Kuncheva, L.: Combining Pattern Classifiers: Methods and Algorithms. Wiley Interscience, Chichester (2004)
Ruta, D., Gabrys, B.: Classifier Selection for Majority Voting. Special issue of the journal of information fusion on Diversity in Multiple Classifier Systems 6(1), 63–81 (2005)
Gabrys, B.: Learning Hybrid Neuro-Fuzzy Classifier Models From Data: To Combine or not to Combine? Fuzzy Sets and Systems 147, 39–56 (2004)
Ruta, D., Gabrys, B.: A Theoretical Analysis of the Limits of Majority Voting Errors for Multiple Classifier Systems. Pattern Analysis and Applications 5, 333–350 (2002)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Gabrys, B., Baruque, B., Corchado, E. (2006). Outlier Resistant PCA Ensembles. In: Gabrys, B., Howlett, R.J., Jain, L.C. (eds) Knowledge-Based Intelligent Information and Engineering Systems. KES 2006. Lecture Notes in Computer Science(), vol 4253. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11893011_55
Download citation
DOI: https://doi.org/10.1007/11893011_55
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-46542-3
Online ISBN: 978-3-540-46544-7
eBook Packages: Computer ScienceComputer Science (R0)