Text Sentiment Classification Algorithm Based on Double Channel Convolutional Neural Network
SHEN Chang1, JI Junzhong1
1.Beijing Municipal Key Laboratory of Multimedia and Intelligent Software Technology, Faculty of Information Technology, Beijing University of Technology, Beijing 100124
Abstract:The existing deep learning method is insufficient to extract features in the text sentiment classification task. To solve the drawback, a text sentiment classification algorithm based on the double channel convolutional neural network with extended features and a dynamic pooling is presented. Firstly, various word features influencing the sentiment orientation of text, such as emotional word, part of speech, adverb of degree, negative word and punctuation, are combined to obtain an extended text feature. Then, the word vector feature and the extended text feature are used as two individual channels of the convolutional neural network, and a new dynamic k-max pooling strategy is adopted to improve the efficiency of feature extraction. The experimental results on standard English datasets demonstrate that the proposed algorithm achieves better classification efficiency than traditional convolutional neural network algorithm with single channel, and it is more advantageous compared with some elitist text sentiment classification algorithms.
[1] 赵妍妍,秦 兵,刘 挺.文本情感分析.软件学报, 2010, 21(8): 1834-1848. (ZHAO Y Y, QIN B, LIU T. Sentiment Analysis. Journal of Software, 2010, 21(8): 1834-1848.) [2] TURNEY P D, LITTMAN M L. Measuring Praise and Criticism: Inference of Semantic Orientation from Association. ACM Transactions on Information Systems, 2003, 21(4): 315-346. [3] HU M Q, LIU B. Mining and Summarizing Customer Reviews // Proc of the 10th ACM SIGKDD International Conference on Know-ledge Discovery and Data Mining. New York, USA: ACM, 2004: 168-177. [4] KIM S M, HOVY E. Automatic Detection of Opinion Bearing Words and Sentences[C/OL]. [2017-10-23]. http://www.aclweb.org/anthology/I05-2011. [5] DING X W, LIU B, YU P S. A Holistic Lexicon-Based Approach to Opinion Mining // Proc of the International Conference on Web Search and Data Mining. New York, USA: 2008: 231-240. [6] PANG B, LEE L, VAITHYANATHAN S. Thumbs up? Sentiment Classification Using Machine Learning Techniques // Proc of the Conference on Empirical Methods in Natural Language Processing. Stroudsburg, USA: ACL, 2002: 79-86. [7] KIM S M, HOVY E. Automatic Identification of pro and con Reasons in Online Reviews // Proc of the COLING/ACL on Main Conference Poster Sessions. Stroudsburg, USA: ACL, 2006: 483-490. [8] CATAL C, NANGIR M. A Sentiment Classification Model Based on Multiple Classifiers. Applied Soft Computing, 2017, 50: 135-141. [9] SOCHER R, PENNINGTON J, HUANG E H, et al. Semi-supervised Recursive Autoencoders for Predicting Sentiment Distributions // Proc of the Conference on Empirical Methods in Natural Language Processing. Stroudsburg, USA: ACL, 2011: 151-161. [10] SOCHER R, PERELYGIN A, WU J Y, et al. Recursive Deep Models for Semantic Compositionality over a Sentiment Treebank // Proc of the Conference on Empirical Methods in Natural Language Processing. Stroudsburg, USA: ACL, 2013: 1631-1642. [11] DONG L, WEI F R, XU K, et al. Adaptive Multi-compositionality for Recursive Neural Network Models. IEEE/ACM Transactions on Audio, Speech and Language Processing, 2016, 24(3): 422-431. [12] IRSOY O, CARDIE C. Opinion Mining with Deep Recurrent Neural Networks // Proc of the Conference on Empirical Methods in Natural Language Processing. Stroudsburg, USA: ACL, 2014: 720-728. [13] WANG X, LIU Y C, SUN C J, et al. Predicting Polarities of Tweets by Composing Word Embeddings with Long Short-Term Memory // Proc of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Confe-rence on Natural Language Processing. Stroudsburg, USA: ACL, 2015: 1343-1353. [14] ZHANG M S, ZHANG Y, VO D T. Gated Neural Networks for Targeted Sentiment Analysis // Proc of the 30th AAAI Conference on Artificial Intelligence. Palo Alto, USA: AAAI Press, 2016: 3087-3093. [15] KIM Y. Convolutional Neural Networks for Sentence Classification // Proc of the Conference on Empirical Methods in Natural Language Processing. Palo Alto, USA: AAAI Press, 2014: 1746-1751. [16] LAI S W, XU L H, LIU K, et al. Recurrent Convolutional Neural Networks for Text Classification // Proc of the 29th AAAI Confe-rence on Artificial Intelligence. Palo Alto, USA: AAAI Press, 2015: 2267-2273. [17] MAJUMMDER N, PORIA S, GWLBUKH A, et al. Deep Lear-ning-Based Document Modeling for Personality Detection from Text. IEEE Intelligent Systems, 2017, 32(2): 74-79. [18] WANG J, YU L C, LAI K R, et al. Dimensional Sentiment Analysis Using a Regional CNN-LSTM Model // Proc of the 54th Annual Meeting of the Association for Computational Linguistics. Stroudsburg, USA: ACL, 2016: 225-230. [19] VOSOUGHI S, VIJAYARAGHAVAN P, ROY D. Tweet2vec: Learning Tweet Embeddings Using Character-Level CNN-LSTM Encoder-Decoder // Proc of the 39th International ACM SIGIR Conference on Research and Development in Information Retrieval. New York, USA: ACM, 2016: 1041-1044. [20] LE CUN Y, BOSER B, DENKER J S, et al. Handwritten Digit Recognition with a Back-Propagation Network // TOURETZKY D S, ed. Advances in Neural Information Processing Systems 2. San Francisco, USA: Morgan Kaufmann Publishers, 1990: 396-404. [21] LECUN Y, BOTTOU L, BENGIO Y, et al. Gradient-Based Learning Applied to Document Recognition. Proceedings of the IEEE, 1998, 86(11): 2278-2324. [22] LI Q Z, SHAH S, FANG R, et al. Tweet Sentiment Analysis by Incorporating Sentiment-Specific Word Embedding and Weighted Text Features // Proc of the IEEE/WIC/ACM International Conference on Web Intelligence. Washington, USA: IEEE, 2016: 568-571. [23] DUMOULIN V, VISIN F. A Guide to Convolution Arithmetic for Deep Learning[J/OL]. [2017-10-23]. https://arxiv.org/pdf/1603.07285.pdf. [24] HE K M, SUN J. Convolutional Neural Networks at Constrained Time Cost // Proc of the IEEE Conference on Computer Vision and Pattern Recognition. Washington, USA: IEEE, 2015: 5353-5360.