Robust Label Distribution Learning from a Perspective of Local Collaboration
XU Suping1,2, SHANG Lin1,2, ZHOU Yujie1,2
1. Department of Computer Science and Technology, Nanjing University, Nanjing 210023 2. State Key Laboratory for Novel Software Technology, Nanjing University, Nanjing 210023
Abstract:In the most of label distribution learning(LDL)algorithms, the correlations among different labels and the overall structure of label distribution are destroyed to a certain extent. Moreover, most existing LDL algorithms mainly focus on improving the predictive performance of label distribution, while ignoring the significance of computational cost and noise robustness in practical applications. To tackle these issues, a local collaborative representation based label distribution learning algorithm (LCR-LDL) is proposed. In LCR-LDL, an unlabeled sample is treated as a collaborative representation of the local dictionary constructed by the neighborhood of the unlabeled sample, and the discriminating information of representation coefficients is utilized to reconstruct the label distribution of unlabeled sample. Experimental results on 15 real-world LDL datasets show that LCR-LDL effectively improves the predictive performance for LDL tasks with a better robustness and low computational cost.
[1] LI J, XU D, GAO W. Removing Label Ambiguity in Learning-Based Visual Saliency Estimation. IEEE Transactions on Image Processing, 2012, 21(4): 1513-1525. [2] XU S P, YANG X B, YU H L, et al. Multi-label Learning with Label-Specific Feature Reduction. Knowledge-Based Systems, 2016, 104: 52-61. [3] GAN W S, LIN J C, CHAO H C, et al. Data Mining in Distributed Environment: A Survey. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 2017, 7(6). DOI: 10.1002/widm.121. [4] GAN W S, LIN J C, FOURNIER-VIGER P, et al. Extracting Recent Weighted-Based Patterns from Uncertain Temporal Databases. Engineering Applications of Artificial Intelligence, 2017, 61: 161-172. [5] GIBAJA E, VENTURA S. Multi-label Learning: A Review of the State of the Art and Ongoing Research. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 2014, 4(6): 411-444. [6] ZHANG M L, LI Y K Y, LIU X, et al. Binary Relevance for Multi-label Learning: An Overview. Frontiers of Computer Science, 2018, 12(2): 191-202. [7] ZHANG M L, ZHOU Z H. A Review on Multi-label Learning Algorithms. IEEE Transactions on Knowledge and Data Engineering, 2014, 26(8): 1819-1837. [8] LYONS M, AKAMATSU S, KAMACHI M, et al. Coding Facial Expressions with Gabor Wavelets // Proc of the 3rd IEEE International Conference on Automatic Face and Gesture Recognition. Washington, USA: IEEE, 1998: 200-205. [9] ZHOU Y, XUE H, GENG X. Emotion Distribution Recognition from Facial Expressions // Proc of the 23rd ACM International Conference on Multimedia. New York, USA: ACM, 2015: 1247-1250. [10] GENG X, HOU P. Pre-release Prediction of Crowd Opinion on Movies by Label Distribution Learning // Proc of the 24th International Conference on Artificial Intelligence. Palo Alto, USA: AAAI Press, 2015: 3511-3517. [11] SHEN W, ZHAO K, GUO Y L, et al. Label Distribution Learning Forests // GUYON I, LUXBURG U V, BENGIO S, et al., eds. Advances in Neural Information Processing Systems 30. Cambridge, USA: The MIT Press, 2017: 834-843. [12] GENG X. Label Distribution Learning. IEEE Transactions on Know-ledge and Data Engineering, 2016, 28(7): 1734-1748. [13] GENG X, JI R. Label Distribution Learning // Proc of the 13th IEEE International Conference on Data Mining Workshops. Wa-shington, USA: IEEE, 2013: 377-383. [14] GENG X, SMITH-MILES K, ZHOU Z H. Facial Age Estimation by Learning from Label Distributions // Proc of the 24th AAAI Conference on Artificial Intelligence. Palo Alto, USA: AAAI Press, 2010: 451-456. [15] GENG X, YIN C, ZHOU Z H. Facial Age Estimation by Learning from Label Distributions. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2013, 35(10): 2401-2412. [16] PIETRA S D, PIETRA V D, LAFFERTY J. Inducing Features of Random Fields. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1997, 19(4): 380-393. [17] NOCEDAL J, WRIGHT S. Numerical Optimization. 2nd Edition. Berlin, Germany: Springer, 2006. [18] XING C, GENG X, XUE H. Logistic Boosting Regression for Label Distribution Learning // Proc of the IEEE Conference on Computer Vision and Pattern Recognition. Washington, USA: IEEE, 2016: 4489-4497. [19] AMALDI F, KANN V. On the Approximability of Minimizing Nonzero Variables or Unsatisfied Relations in Linear Systems. Theoretical Computer Science, 1998, 209(1/2): 237-260. [20] HUANG S, YANG Y, YANG D, et al. Class Specific Sparse Re-presentation for Classification. Signal Processing, 2015, 116: 38-42. [21] LU C Y, MIN H, GUI J, et al. Face Recognition via Weighted Sparse Representation. Journal of Visual Communication and Image Representation, 2013, 24(2): 111-116. [22] WRIGHT J, YANG A Y, GANESH A, et al. Robust Face Recognition via Sparse Representation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2009, 31(2): 210-227. [23] ZHANG L, YANG M, FENG X C. Sparse Representation or Co-llaborative Representation: Which Helps Face Recognition? // Proc of the International Conference on Computer Vision. Washington, USA: IEEE, 2011: 471-478. [24] SHAO C B, SONG X N, FENG Z H, et al. Dynamic Dictionary Optimization for Sparse-Representation-Based Face Classification Using Local Difference Images. Information Sciences, 2017, 393: 1-14. [25] FRIEDMAN M. A Comparison of Alternative Tests of Significance for the Problem of m Rankings. Annals of Mathematical Statistics, 1940, 11(1): 86-92. [26] DEMSAR J. Statistical Comparisons of Classifiers over Multiple Data Sets. Journal of Machine Learning Research, 2016, 7: 1-30.