Abstract
Since the long short-term memory (LSTM) network is a sequential structure, it is difficult to effectively represent the structural level information of the context. Sentiment analysis based on the original LSTM causes a problem of structural level information loss, and its capacity to capture the context information is finite. To address this problem, we proposed a novel structure-attention-based LSTM as a hierarchical structure model. It may capture relevant information in the context as much as possible. We propose HM (ht matrix) to storage the structural information of the context. Furthermore, we introduce the attention mechanism to realize vector selection. Compared with the original LSTM and normal attention-based sentiment classification methods, our model obtains a higher classification precision. It is proved that the structure-attention-based method proposed in this study has an advantage in capturing the potential semantic structure.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Pang, B., Lee, L., Vaithyanathan, S.: Sentiment classification using machine learning techniques. In: Conference on Empirical Methods in Natural Language Processing, vol. 10, pp. 79–86. ACM, Stroudsburg, PA, USA (2002)
Pang, B., Lee, L.: Opinion mining and sentiment analysis. Found. Trends Inf. Retr. 2(1–2), 1–135 (2008)
Yu, H., Hatzivassiloglou, V.: Towards answering opinion questions: separating facts from opinions and identifying the polarity of opinion sentences. In: EMNLP, pp. 129–136 (2003)
Khoo, H., Zhou, Na, J.C., et al.: Effectiveness of simple linguistic processing in automatic sentiment classification of product reviews. In: ISKO, pp. 49–54. Ergon Verlag, Wurzburg, Germany (2004)
Bengio, Y., Ducharme, R., Vincent, P., et al.: A neural probabilistic language model. J. Mach. Learn. Res. 3, 1137–1155 (2003)
Gers, F.A., Schmidhuber, J., Cummins, F.: Learning to forget: continual prediction with LSTM. In: 9th International Conference on IET, pp. 2002–2451 (1999)
Stollenga, M.F., Byeon, W., Liwicki, M., et al.: Parallel multi-dimensional LSTM, with application to fast biomedical volumetric image segmentation. Comput. Sci. (2015)
Tang, D., Qin, B., Liu, T.: Document modeling with gated recurrent neural network for sentiment classification. In: EMNLP, pp. 1422–1432 (2005)
Mnih, V., Heess, N., Graves, A.: Recurrent models of visual attention. In: Advances in Neural Information Processing Systems, 2014, pp. 2204–2212 (2014)
Bahdanau, D., Cho, K., Bengio, Y.: Neural Machine Translation by Jointly Learning to Align and Translate. In: ICLR 2015, pp. 1–15 (2015)
Luong, M.T., Pham, H., Manning, C.D.: Effective approaches to attention-based neural machine translation. Comput. Sci. (2015)
Kokkinos, F., Potamianos, A.: Structural attention neural networks for improved sentiment analysis (2017)
Acknowledgement
This work is supported by the Nature Science Foundation of China (No. 61402386, No. 61305061, No. 61502105, No. 61572409, No. 81230087 and No. 61571188), Open Fund Project of Fujian Provincial Key Laboratory of Information Processing and Intelligent Control (Minjiang University) (No. MJUKF201743), and Education and scientific research projects of young and middle-aged teachers in Fujian Province under Grand No. JA15075. Fujian Province 2011 Collaborative Innovation Center of TCM Health Management and Collaborative Innovation Center of Chinese Oolong Tea Industry — Collaborative Innovation Center (2011) of Fujian Province.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG
About this paper
Cite this paper
Lin, K., Lin, D., Cao, D. (2018). Sentiment Analysis Model Based on Structure Attention Mechanism. In: Chao, F., Schockaert, S., Zhang, Q. (eds) Advances in Computational Intelligence Systems. UKCI 2017. Advances in Intelligent Systems and Computing, vol 650. Springer, Cham. https://doi.org/10.1007/978-3-319-66939-7_2
Download citation
DOI: https://doi.org/10.1007/978-3-319-66939-7_2
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-66938-0
Online ISBN: 978-3-319-66939-7
eBook Packages: EngineeringEngineering (R0)