Abstract:In most of the existing manifold learning algorithms, the geometry structure of the data instances is preserved, but the label information is ignored. Therefore, the application of manifold learning algorithms in data classification is limited. In this paper, a semi-supervised manifold learning algorithm based on neighborhood components analysis is proposed. A distance metric matrix is learned by using neighbor components analysis and local neighbors of the sample points are selected by using the new distance metric. The local geometric structures of the sample points and their neighbors are constructed under the new distance metric, and the local geometric structures are preserved in the low-dimensional embedding coordinates of the sample points. The classification experiments conducted on three different datasets demonstrate the efficiency of the proposed algorithm.
[1] TENENBAUM J B, DE SILVA V, LANGFORD J C. A Global Geometric Framework for Nonlinear Dimensionality Reduction. Science, 2000, 290(5500): 2319-2323. [2] ROWEIS S T, SAUL L K. Nonlinear Dimensionality Reduction by Locally Linear Embedding. Science, 2000, 290(5500): 2323-2326. [3] BELKIN M, NIYOGI P. Laplacian Eigenmaps for Dimensionality Reduction and Data Representation. Neural Computation, 2003, 15(6): 1373-1396. [4] ZHANG Z Y, ZHA H Y. Principal Manifolds and Nonlinear Dimension Reduction via Local Tangent Space Alignment. SIAM Journal of Scientific Computing, 2005, 26(1): 313-338. [5] DE RIDDER D, DUIN R P W. Locally Linear Embedding for Classification. Technical Report, PH-2002-01. Delft, The Netherlands: Delft University of Technology, 2002. [6] CHEN M X, WANG J. Semi-supervised Learning on Multi-manifold. Journal of Computational Information Systems, 2014, 10(12): 5131-5138. [7] GOLDBERGER J, ROWEIS S T, HINTON G E, et al. Neighbourhood Components Analysis[J/OL]. [2017-01-13]. http://www.cs.toronto.edu/~hinton/absps/nca.pdf. [8] WEINBERGER K Q, SAUL L K. Distance Metric Learning for Large Margin Nearest Neighbor Classification. Journal of Machine Learning Research, 2009, 10: 207-244. [9] DAVIS J V, KULIS B, JAIN P, et al. Information-Theoretic Metric Learning // Proc of the 24th International Conference on Machine Learning. New York, USA: ACM, 2007: 209-216. [10] GUILLAUMIN M, VERBEEK J, SCHMID C. Is That You? Metric Learning Approaches for Face Identification // Proc of the 12th IEEE International Conference on Computer Vision. Washington, USA: IEEE, 2009: 498-505. [11] NGUYEN H V, BAI L. Cosine Similarity Metric Learning for Face Verification // Proc of the 10th Asian Conference on Computer Vision. Berlin, Germany: Springer, 2010, II: 709-720. [12] ELHAMIFAR E, VIDAL R. Sparse Manifold Clustering and Embedding // SHAWE-TAYLOR J, ZEMEL R S, BARTLETT P L, et al., eds. Advances in Neural Information Processing Systems 24. Cambridge, USA: The MIT Press, 2011: 55-63. [13] WANG J. Real Local-Linearity Preserving Embedding. Neurocomputing, 2014, 136: 7-13. [14] WANG J, ZHANG Z Y. Nonlinear Embedding Preserving Multiple Local-Linearities. Pattern Recognition, 2010, 43(4): 1257-1268. [15] SIGILLITO V G, WING S P, HUTTON L V, et al. Classification of Radar Returns from the Ionosphere Using Neural Networks. Johns Hopkins APL Technical Digest, 1989, 10(3): 262-266. [16] SIM T, BAKER S, BSAT M. The CMU Pose, Illumination, and Expression(PIE) Database // Proc of the 5th IEEE International Conference on Automatic Face and Gesture Recognition. Washington, USA: IEEE, 2002: 46-51. [17] HULL J J. A Database for Handwritten Text Recognition Research. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1994, 16(5): 550-554.