Abstract
The fundamental assumption that training and operational data come from the same probability distribution, which is the basis of most learning algorithms, is often not satisfied in practice. Several algorithms have been proposed to cope with classification problems where the class priors may change after training, but they can show a poor performance when the class conditional data densities also change. In this paper, we propose a re-estimation algorithm that makes use of unlabeled operational data to adapt the classifier behavior to changing scenarios. We assume that (a) the classes may be decomposed in several (unknown) subclasses, and (b) the prior subclass probabilities may change after training. Experimental results with practical applications show an improvement over an adaptive method based on class priors, while preserving a similar performance when there are no subclass changes.
Supported by the Spanish MEC projects DPI2006-02550 and TEC2008-01348/TEC.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Alaiz-Rodríguez, R., Guerrero-Curieses, A., Cid-Sueiro, J.: Minimax regret classifier for imprecise class distributions. Journal of Machine Learning Research 8, 103–130 (2007)
Alba, J., Docio, L., Docampo, D., Marquez, O.: Growing gaussian mixtures network for classification applications. Signal Process 76(1), 43–60 (1999)
Arribas, J.I., Cid-Sueiro, J.: A model selection algorithm for a posteriori probability estimation with neural networks. IEEE Transactions on Neural Networks 16(4), 799–809 (2005)
Flawcett, T., Flach, P.: A response to webb and ting’s on the application of roc analysis to predict classification performance under varying class distributions. Machine Learning 58, 33–38 (2005)
Holte, R.: Elaboration on two points raised in classifier technology and the illusion of progress. Statistical Science 21(1) (2006)
Provost, F., Fawcett, T.: Robust classification systems for imprecise environments. Machine Learning 42(3), 203–231 (2001)
Saerens, M., Latinne, P., Decaestecker, C.: Adjusting a classifier for new a priori probabilities: A simple procedure. Neural Computation 14, 21–41 (2002)
Shimodaira, H.: Improving predictive inference under convariance shift by weighting the log-likelihood function. Journal of Statistical Planning and Inference (2000)
Webb, G., Ting, K.: On the application of roc analysis to predict classification performance under varying class distributions. Machine Learning 58(1), 25–32 (2005)
Yamazaki, K., Kawanabe, M., Watanabe, S., Sugiyama, M., Müller, K.: Asymptotic bayesian generalization error when training and test distributions are different. In: Proc. of the 24th international conference on Machine learning (2007)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2009 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Alaiz-Rodríguez, R., Guerrero-Curieses, A., Cid-Sueiro, J. (2009). Improving Classification under Changes in Class and Within-Class Distributions. In: Cabestany, J., Sandoval, F., Prieto, A., Corchado, J.M. (eds) Bio-Inspired Systems: Computational and Ambient Intelligence. IWANN 2009. Lecture Notes in Computer Science, vol 5517. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-02478-8_16
Download citation
DOI: https://doi.org/10.1007/978-3-642-02478-8_16
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-02477-1
Online ISBN: 978-3-642-02478-8
eBook Packages: Computer ScienceComputer Science (R0)