Abstract
High Dimensional Data (HDD) is one of the biggest challenges in Data Science arising from Big Data. The application of dimensionality reduction techniques over HDD allows visualization and, thus, a better problem understanding. In addition, these techniques also can enhance the performance of Machine Learning (ML) algorithms while increasing the explanatory power. This paper presents an automatic method capable of obtaining an adequate representation of the data, given a previously trained ML model. Likewise, an automatic method is introduced to bring a Support Vector Machine (SVM) model based on an adequate representation of the data. Both methods provide an Explanaible Machine Learning procedure. The proposal is tested on several data sets providing promising results. It significantly eases the visualization and understanding task to the data scientist when a ML model has already been trained, as well as the ML selection parameters when a reduced representation of data has been achieved.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Amaratunga, D., Cabrera, J.: High-dimensional data. Journal of the National Science Foundation of Sri Lanka 44(1) (2016)
Arrieta, A.B., et al.: Explainable artificial intelligence (xai): Concepts, taxonomies, opportunities and challenges toward responsible ai. Information Fusion 58, 82–115 (2020)
Ayesha, S., Hanif, M.K., Talib, R.: Overview and comparative study of dimensionality reduction techniques for high dimensional data. Information Fusion 59, 44–58 (2020)
Breiman, L., Friedman, J.H., Olshen, R.A., Stone, C.J.: Classification and regression trees. Routledge (2017)
Dua, D., Graff, C.: UCI machine learning repository (2017), http://archive.ics.uci.edu/ml
Han, S., Qubo, C., Meng, H.: Parameter selection in svm with rbf kernel function. In: World Automation Congress 2012. pp. 1–4. IEEE (2012)
Higham, N.J.: The symmetric procrustes problem. BIT Numer. Math. 28(1), 133–143 (1988)
Van der Maaten, L., Hinton, G.: Visualizing data using t-sne. J. Mach. Learn. Res. 9(11), 2579–2605 (2008)
McInnes, L., Healy, J., Melville, J.: Umap: Uniform manifold approximation and projection for dimension reduction. arXiv preprint arXiv:1802.03426 (2018)
Moguerza, J.M., Muñoz, A.: Support vector machines with applications. Stat. Sci. 21(3), 322–336 (2006)
Torgerson, W.S.: Multidimensional scaling: i. theory and method. Psychometrika 17(4), 401–419 (1952) https://doi.org/10.1007/BF02288916
Vert, J.P., Tsuda, K., Schölkopf, B.: A primer on kernel methods. Kernel Methods Comput. Biol. 47, 35–70 (2004)
Acknowledgements
This research has been supported by grants from Rey Juan Carlos University (Ref: C1PREDOC2020), Madrid Autonomous Community (Ref: IND2019/TIC-17194) and the Spanish Ministry of Economy and Competitiveness, under the Retos-Investigación program: MODAS-IN (Ref: RTI-2018-094269-B-I00).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Cuesta, M., Martín de Diego, I., Lancho, C., Aceña, V., M. Moguerza, J. (2021). From Classification to Visualization: A Two Way Trip. In: Yin, H., et al. Intelligent Data Engineering and Automated Learning – IDEAL 2021. IDEAL 2021. Lecture Notes in Computer Science(), vol 13113. Springer, Cham. https://doi.org/10.1007/978-3-030-91608-4_29
Download citation
DOI: https://doi.org/10.1007/978-3-030-91608-4_29
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-91607-7
Online ISBN: 978-3-030-91608-4
eBook Packages: Computer ScienceComputer Science (R0)