Abstract
Least squares regression is an effective multi-classification method; however, in practical applications, many models based on the least squares regression method are significantly affected by noise (and outliers). Therefore, effectively reducing the adverse effects of noise is conducive to obtaining a better classification performance. Besides, preserving the intrinsic characteristics of samples to the greatest extent possible is beneficial for improving the discriminative ability of the model. Based on this analysis, we propose the relaxed group low-rank regression model for multi-class classification. The model effectively captures the hidden structural information of samples by exploiting the group low-rank constraint. Meanwhile, with the group low-rank constraint and the graph embedding constraint, the proposed method has more tolerance to noise (and outliers). The feature matrix with the L21-norm and the graph embedding constraint complement each other to capture the intrinsic characteristics of the samples. In addition, a sparsity error term with the L21 norm is utilized to relax the strict target label matrix. These factors guarantee that the original samples are converted into a more compact and discriminative characteristic space. Finally, we compare the proposed model with various popular algorithms on several benchmark datasets. The experimental results demonstrate that the performance of the proposed method outperforms those of state-of-the-art methods.
Similar content being viewed by others
References
Asgharian L, Ebrahimnezhad H (2020) How many sample points are sufficient for 3D model surface representation and accurate mesh simplification? Multimed Tools Appl 79:29595–29620. https://doi.org/10.1007/s11042-020-09395-3
Cai J, Candes E, Shen Z (2010) A singular value Thresholding algorithm for matrix completion. SIAM J Optim 20(4):1956–1982
Cai X, Ding C, Nie F, Huang H (2013) On the equivalent of low-rank linear regressions and linear discriminant analysis based regressions. In: Proceedings of the ACM SIGKDD international conference on knowledge discovery and data mining, pp. 1124–1132
Chang K, Hsieh C, Lin C (2008) Coordinate descent method for large-scale L2-loss linear support vector machines. J Mach Learn Res 9:1369–1398
Chen Y, Xu W, Zuo J, Yang K (2019) The fire recognition algorithm using dynamic feature fusion and IV-SVM classifier. Clust Comput 22:7665–7675
Chen Y, Xiong J, Xu W, Zuo J (2019) A novel online incremental and decremental learning algorithm based on variable support vector machine. Clust Comput 22:7435–7445
Chen Y, Wang J, Liu S, Chen X, Xiong J, Xie J, Yang K (2019) Multiscale fast correlation filtering tracking algorithm based on a feature fusion model. Concurrency and Computation-Practice and Experience. https://doi.org/10.1002/cpe.5533
Chen Y, Wang J, Xia R, Zhang Q, Cao Z, Yang K (2019) The visual object tracking algorithm research based on adaptive combination kernel. J Ambient Intell Humaniz Comput 10(12):4855–4867
Chen Y, Wang J, Chen X, Zhu M, Yang K, Wang Z, Xia R (2019) Single-image super-resolution algorithm based on structural self-similarity and deformation block features. IEEE Access 7:58791–58801
Chen Y, Tao J, Zhang Q, Yang K, Chen X, Xiong J, Xia R, Xie J (2020) Saliency detection via the improved hierarchical principal component analysis method. Wirel Commun Mob Comput 2020:1–12. https://doi.org/10.1155/2020/8822777
Chen Y, Tao J, Liu L, Xiong J, Xia R, Xie J, Zhang Q, Yang K (2020) Research of improving semantic image segmentation based on a feature fusion model. J Ambient Intell Humaniz Comput. https://doi.org/10.1007/s12652-020-02066-z
He X, Niyogi P (2003) Locality preserving projections, In: Proceedings of Conference on Advances in Neural Information Processing Systems (NIPS), pp 234–241
Hosmer D, Lemeshow J, Sturdivant R (2013) Applied logistic regression. Wiley, New York
Li Y, Ngom A (2013) Nonnegative least-squares methods for the classification of high-dimensional biological data. IEEE/ACM trans. Comput Biol Bioinf 10(2):447–456
Lin Z, Chen M, Ma Y (2010) The augmented Lagrange multiplier method for exact recovery of corrupted low-rank matrice. arXiv preprint arXiv:1009.5055
Lu X, Ma C, Ni B, et al (2018) Deep regression tracking with shrinkage loss. In: European conference on computer vision, pp. 369–386
Lu X, Ma C, Ni B, Yang X (2019) Adaptive region proposal with channel regularization for robust object tracking. IEEE Transactions on Circuits and Systems for Video Technology:1. https://doi.org/10.1109/TCSVT.2019.2944654
Nene S, Nayar S, Murase H (1996) Columbia object image library (COIL-100). Technical report CUCS-006-96
Nie F, Huang H, Cai X, Ding C (2010) Efficient and robust feature selection via joint L2,1-norms minimization, In: Proc. Adv. Neural Inf. Process. Syst. pp 1813–1821
Nie F, Xiang S, Liu Y, Hou C, Zhang C (2012) Orthogonal vs. uncorrelated least squares discriminant analysis for feature extraction. Pattern Recognit. Lett. 33(5):485–491
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition arXiv preprint arXiv: 1409.1556
Wang L, Pan C (2018) Groupwise retargeted least-squares regression. IEEE Trans Neural Netw 29(4):1352–1358
Wei L, Wang X, Wu A, Zhou R, Zhu C (2018) Robust subspace segmentation by self-representation constrained low-rank representation. Neural Process Lett 48:1671–1691. https://doi.org/10.1007/s11063-018-9783-y
Wen J, Xu Y, Li Z, Ma Z, Xu Y (2018) Inter-class sparsity based discriminative least square regression. Neural Netw 102:36–47
Wright J, Yang A, Ganesh A, Sastry S, Ma Y (2009) Robust face recognition via sparse representation. IEEE Trans Pattern Anal Mach Intell 31(2):210–227
Xiang S, Zhu Y, Shen X, Ye J (2012) Optimal exact least squares rank minimization. In: Proceedings of the ACM SIGKDD international conference on knowledge discovery and data mining, pp. 480–488
Xiang S, Nie F, Meng G, Pan C, Zhang C (2012) Discriminative least squares regression for multiclass classification and feature selection. IEEE Trans Neural Netw 23(11):1738–1754
Xu Y, Fang X, Zhu Q, Chen Y, You J, Liu H (2014) Modified minimum squared error algorithm for robust classification and face recognition experiments. Neurocomputing 135:253–261
Xu Y, Fang X, Wu J, Li X, Zhang D (2016) Discriminative transfer subspace learning via low-rank and sparse representation. IEEE Trans Image Process 25(2):850–863
Yuan H, Zheng J, Lai L et al (2018) A constrained least squares regression model. Inf Sci 429:247–259
Zhang L, Yang M, Feng X (2011) Sparse representation or collaborative representation: which helps face recognition? In: IEEE international conference on computer vision, Barcelona, Spain, pp. 471–478
Zhang X, Wang L, Xiang S, Liu C (2015) Retargeted least squares regression algorithm. IEEE Trans Neural Netw 26(9):2206–2213
Zhang Z, Xu Y, Yang J, Li X, Zhang D (2015) A survey of sparse representation: algorithms and applications. IEEE Access 3:490–530
Zhang Z, Lai Z, Xu Y, Shao L, Wu J, Xie GS (2017) Discriminative elastic-net regularized linear regression. IEEE Trans Image Process 26(3):1466–1481
Zhao H, Wang Z, Nie F (2016) Orthogonal least squares regression for feature extraction. Neurocomputing 216:200–207
Zheng W, Xin M, Wang X, Wang B (2014) A novel speech emotion recognition method via incomplete sparse least square regression. Signal process. Lett. https://doi.org/10.1109/LSP.2014.2308954. 1–1
Acknowledgments
This paper is supported by the Graduate Innovation Foundation of Jiangsu Province under Grant No. KYLX16_0781, the Natural Science Foundation of Jiangsu Province under Grants No. BK20181340, the 111 Project under Grants No. B12018, and PAPD of Jiangsu Higher Education Institutions.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Wang, S., Ge, H., Yang, J. et al. Relaxed group low rank regression model for multi-class classification. Multimed Tools Appl 80, 9459–9477 (2021). https://doi.org/10.1007/s11042-020-10080-8
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-020-10080-8