Skip to main content

Application of Deep Learning Approaches for Sentiment Analysis

  • Chapter
  • First Online:
Deep Learning-Based Approaches for Sentiment Analysis

Abstract

Social media platforms, forums, blogs, and opinion sites generate vast amount of data. Such data in the form of opinions, emotions, and views about services, politics, and products are characterized by unstructured format. End users, business industries, and politicians are highly influenced by sentiments of the people expressed on social media platforms. Therefore, extracting, analyzing, summarizing, and predicting the sentiments from large unstructured data needs automated sentiment analysis. Sentiment analysis is an automated process of extracting the opinionated from data and classifying the sentiments as positive, negative, and neutral. Lack of enough labeled data for sentiment analysis is one of the crucial challenges in natural language processing. Deep learning has emanated as one of the highly sought-after solutions to address this challenge due to automated and hierarchical learning capability inherently supported by deep learning models. Considering the application of deep learning approaches for sentiment analysis, this chapter aims to put forth taxonomy of traits to be considered for deep learning-based sentiment analysis and demystify the role of deep learning approaches for sentiment analysis.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 139.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 179.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 179.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Pathak, A.R., M. Pandey, and S. Rautaray. 2018. Construing the big data based on taxonomy, analytics and approaches. Iran Journal of Computer Science 1: 237–259.

    Article  Google Scholar 

  2. NLP market. https://www.tractica.com/newsroom/press-releases/natural-language-processing-market-to-reach-22–3-billion-by-2025/.

  3. Agarwal, B., and N. Mittal. 2016. Prominent feature extraction for sentiment analysis. In Springer Book Series: Socio-Affective Computing series, 1–115. Springer International Publishing, ISBN: 978-3-319-25343-5. https://doi.org/10.1007/978-3-319-25343-5.

    Book  Google Scholar 

  4. Pathak, A.R., M. Pandey, and S. Rautaray. 2018. Application of deep learning for object detection. Procedia Computer Science 132: 1706–1717.

    Article  Google Scholar 

  5. Pathak, A.R., M. Pandey, and S. Rautaray. 2018. Deep learning approaches for detecting objects from images: A review. In Progress in Computing, Analytics and Networking, ed. Pattnaik, P. K., S.S. Rautaray, H. Das, J. Nayak, J., 491–499. Springer Singapore.

    Google Scholar 

  6. Pathak, A.R., M. Pandey, S. Rautaray, and K. Pawar. 2018. Assessment of object detection using deep convolutional neural networks. In Advances in Intelligent Systems and Computing, 673.

    Google Scholar 

  7. Pawar, K., and V. Attar. 2019. Deep learning approaches for video-based anomalous activity detection. World Wide Web 22: 571–601.

    Article  Google Scholar 

  8. Pawar, K., and V. Attar. 2019. Deep Learning approach for detection of anomalous activities from surveillance videos. In CCIS. Springer, In Press.

    Google Scholar 

  9. Pathak, A.R., M. Pandey, and S. Rautaray. 2019. Adaptive model for dynamic and temporal topic modeling from big data using deep learning architecture. Internationl Journal of Intelligent Systems and Applications 11: 13–27. https://doi.org/10.5815/ijisa.2019.06.02.

    Article  Google Scholar 

  10. Bhat, M.R., M.A. Kundroo, T.A. Tarray, and B. Agarwal. 2019. Deep LDA: A new way to topic model. Journal of Information and Optimization Sciences 1–12 (2019).

    Google Scholar 

  11. Pathak, A.R., M. Pandey, and S. Rautaray. 2019. Adaptive framework for deep learning based dynamic and temporal topic modeling from big data. Recent Patents on Engineering, Bentham Science 13: 1. https://doi.org/10.2174/1872212113666190329234812.

  12. Pathak, A.R., M. Pandey, and S. Rautaray. 2019. Empirical evaluation of deep learning models for sentiment analysis. Journal of Statistics and Management Systems 22: 741–752.

    Article  Google Scholar 

  13. Pathak, A.R., M. Pandey, and S. Rautaray. 2019. Adaptive model for sentiment analysis of social media data using deep learning. In International Conference on Intelligent Computing and Communication Technologies, 416–423.

    Google Scholar 

  14. Ram, S., S. Gupta, and B. Agarwal. 2018. Devanagri character recognition model using deep convolution neural network. Journal of Statistics and Management Systems, 21: 593–599.

    Article  Google Scholar 

  15. Hinton, G., and et al. 2012. Deep neural networks for acoustic modeling in speech recognition. IEEE Signal Processing Magazine, 29.

    Google Scholar 

  16. Jain, G., M. Sharma, and B. Agarwal. 2019. Spam detection in social media using convolutional and long short term memory neural network. Annals of Mathematics and Artificial Intelligence 85: 21–44.

    Article  Google Scholar 

  17. Agarwal, B., H. Ramampiaro, H. Langseth, and M. Ruocco. 2018. A deep network model for paraphrase detection in short text messages. Information Processing & Management 54: 922–937.

    Article  Google Scholar 

  18. Jain, G., M. Sharma, and B. Agarwal. 2019. Optimizing semantic LSTM for spam detection. International Journal of Information Technology 11: 239–250.

    Article  Google Scholar 

  19. Liu, B. 2012. Sentiment Analysis and Opinion Mining, 1–108. https://doi.org/10.2200/s00416ed1v01y201204hlt016.

    Article  Google Scholar 

  20. SemEval-2014. http://alt.qcri.org/semeval2014/task4/.

  21. Glorot, X., A. Bordes, and Y. Bengio. 2011. Domain adaptation for large-scale sentiment classification: A deep learning approach. In Proceedings of the 28th International Conference on Machine Learning (ICML-11), 513–520.

    Google Scholar 

  22. Bengio, Y., R. Ducharme, P. Vincent, and C. Jauvin. 2003. A neural probabilistic language model. Journal of Machine Learning Research 3: 1137–1155.

    MATH  Google Scholar 

  23. Collobert, R., et al. 2011. Natural language processing (almost) from scratch. Journal of Machine Learning Research 12: 2493–2537.

    MATH  Google Scholar 

  24. Mikolov, T., K. Chen, G. Corrado, and J. Dean. 2013. Efficient estimation of word representations in vector space. arXiv Prepr. arXiv1301.3781.

    Google Scholar 

  25. Bojanowski, P., E. Grave, A. Joulin, and T. Mikolov. 2017. Enriching word vectors with subword information. Transactions of the Association for Computational Linguistics 5: 135–146.

    Article  Google Scholar 

  26. Pennington, J., R. Socher, and C. Manning. 2014. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), 1532–1543.

    Google Scholar 

  27. Peters, M.E., and et al. 2018. Deep contextualized word representations. In Proceedings of NAACL.

    Google Scholar 

  28. Tang, D., and et al. 2014. Learning sentiment-specific word embedding for twitter sentiment classification. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 1555–1565.

    Google Scholar 

  29. Yang, Z., and et al. 2018. Glomo: Unsupervisedly learned relational graphs as transferable representations. arXiv Prepr. arXiv1806.05662.

    Google Scholar 

  30. Howard, J., and S. Ruder. Universal language model fine-tuning for text classification. arXiv Prepr. arXiv1801.06146.

    Google Scholar 

  31. Radford, A., K. Narasimhan, T. Salimans, I. Sutskever. 2018. Improving language understanding by generative pre-training. URL https://s3-us-west-2.amazonaws.com/openai-assets/research-covers/language-unsupervised/language_understanding_paper.pdf.

  32. Devlin, J., M.-W. Chang, K. Lee, and K. Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv Prepr. arXiv1810.04805.

    Google Scholar 

  33. Liu, P., S. Joty, and H. Meng. 2015. Fine-grained opinion mining with recurrent neural networks and word embeddings. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, 1433–1443.

    Google Scholar 

  34. Zhai, S., and Z.M. Zhang. 2016. Semisupervised autoencoder for sentiment analysis. In Thirtieth AAAI Conference on Artificial Intelligence.

    Google Scholar 

  35. EMLo. https://allennlp.org/elmo.

  36. Zhu, Y., and et al. 2015. Aligning books and movies: Towards story-like visual explanations by watching movies and reading books. In Proceedings of the IEEE International Conference on Computer Vision, 19–27.

    Google Scholar 

  37. Poria, S., E. Cambria, and A. Gelbukh. 2016. Aspect extraction for opinion mining with a deep convolutional neural network. Knowledge-Based Systems 108: 42–49.

    Article  Google Scholar 

  38. Wang, W., S.J. Pan, D. Dahlmeier, and X. Xiao. 2016. Recursive neural conditional random fields for aspect-based sentiment analysis. arXiv Prepr. arXiv1603.06679.

    Google Scholar 

  39. Jebbara, S., and P. Cimiano. 2016. Aspect-based relational sentiment analysis using a stacked neural network architecture. In Proceedings of the Twenty-second European Conference on Artificial Intelligence, 1123–1131.

    Google Scholar 

  40. Tang, D., B. Qin, and T. Liu. 2015. Learning semantic representations of users and products for document level sentiment classification. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), 1014–1023.

    Google Scholar 

  41. Yu, J., and J. Jiang. 2016. Learning sentence embeddings with auxiliary tasks for cross-domain sentiment classification. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, 236–246.

    Google Scholar 

  42. Sarma, P.K., Y. Liang, and W.A. Sethares. 2018. Domain adapted word embeddings for improved sentiment classification. arXiv Prepr. arXiv1805.04576.

    Google Scholar 

  43. Bengio, Y., J.-S. Senécal, and Others. 2003. Quick training of probabilistic neural nets by importance sampling. In AISTATS, 1–9.

    Google Scholar 

  44. Bengio, Y., and J.-S. Senécal. 2008. Adaptive importance sampling to accelerate training of a neural probabilistic language model. IEEE Transactions on Neural Networks 19: 713–722.

    Article  Google Scholar 

  45. Jean, S., K. Cho, R. Memisevic, and Y. Bengio. 2014. On using very large target vocabulary for neural machine translation. arXiv Prepr. arXiv1412.2007.

    Google Scholar 

  46. Mnih, A., and Y.W. Teh. 2012. A fast and simple algorithm for training neural probabilistic language models. arXiv Prepr. arXiv1206.6426.

    Google Scholar 

  47. Morin, F., and Y. Bengio. 2005. Hierarchical probabilistic neural network language model. Aistats 5: 246–252.

    Google Scholar 

  48. Chen, W., D. Grangier, and M. Auli. 2015. Strategies for training large vocabulary neural language models. arXiv Prepr. arXiv1512.04906.

    Google Scholar 

  49. Kim, Y., Y. Jernite, D. Sontag, and A.M. Rush. 2016. Character-aware neural language models. In Thirtieth AAAI Conference on Artificial Intelligence.

    Google Scholar 

  50. Jozefowicz, R., O. Vinyals, M. Schuster, N. Shazeer, and Y. Wu. 2016. Exploring the limits of language modeling. arXiv Prepr. arXiv1602.02410.

    Google Scholar 

  51. Tang, D., B. Qin, and T. Liu. 2015. Document modeling with gated recurrent neural network for sentiment classification. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, 1422–1432.

    Google Scholar 

  52. Zhou, H., L. Chen, F. Shi, and D. Huang. 2015. Learning bilingual sentiment word embeddings for cross-language sentiment classification. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), 430–440.

    Google Scholar 

  53. Dou, Z.-Y. 2017. Capturing user and product information for document level sentiment analysis with deep memory network. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, 521–526.

    Google Scholar 

  54. Le, Q., and T. Mikolov. 2014. Distributed representations of sentences and documents. In International Conference on Machine Learning, 1188–1196.

    Google Scholar 

  55. Yang, Z., and et al. 2016. Hierarchical attention networks for document classification. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 1480–1489.

    Google Scholar 

  56. Yin, Y., Y. Song, and M. Zhang. 2017. Document-level multi-aspect sentiment classification as machine comprehension. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, 2044–2054.

    Google Scholar 

  57. Zhou, X., X. Wan, and J. Xiao. 2016. Attention-based LSTM network for cross-lingual sentiment classification. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, 247–256.

    Google Scholar 

  58. Li, Z., Y. Zhang, Y. Wei, Y. Wu, and Q. Yang. 2017. End-to-end adversarial memory network for cross-domain sentiment classification. In IJCAI, 2237–2243.

    Google Scholar 

  59. Rao, G., W. Huang, Z. Feng, and Q. Cong. 2018. LSTM with sentence representations for document-level sentiment classification. Neurocomputing 308: 49–57.

    Article  Google Scholar 

  60. Socher, R., J. Pennington, E.H. Huang, A.Y. Ng, and C.D. Manning. semi-supervised recursive autoencoders for predicting sentiment distributions. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, 151–161.

    Google Scholar 

  61. Socher, R., B. Huval, C.D. Manning, and A.Y. Ng. 2012. Semantic compositionality through recursive matrix-vector spaces. In Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, 1201–1211.

    Google Scholar 

  62. Qian, Q., and et al. 2015. Learning tag embeddings and tag-specific composition functions in recursive neural network. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), 1365–1374.

    Google Scholar 

  63. Kalchbrenner, N., E. Grefenstette, and P. Blunsom. 2014. A convolutional neural network for modelling sentences. arXiv Prepr. arXiv1404.2188.

    Google Scholar 

  64. dos Santos, C., and M. Gatti. 2014. Deep convolutional neural networks for sentiment analysis of short texts. In Proceedings of COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers, 69–78.

    Google Scholar 

  65. Wang, X., Y. Liu, S.U.N. Chengjie, B. Wang, and X. Wang. 2015. Predicting polarities of tweets by composing word embeddings with long short-term memory. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), vol. 1, 1343–1353.

    Google Scholar 

  66. Wang, J., L.-C. Yu, K. Lai, and X. Zhang. 2016. Dimensional sentiment analysis using a regional CNN-LSTM model. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 225–230.

    Google Scholar 

  67. Ruder, S., P. Ghaffari, and J.G. Breslin. 2016. A hierarchical model of reviews for aspect-based sentiment analysis. arXiv Prepr. arXiv1609.02745.

    Google Scholar 

  68. Wang, W., S.J. Pan, and D. Dahlmeier, and X. Xiao. 2016. Recursive neural conditional random fields for aspect-based sentiment analysis. arXiv Prepr. arXiv1603.06679.

    Google Scholar 

  69. Xu, H., B. Liu, L. Shu, and P.S. Yu. 2018. Double embeddings and cnn-based sequence labeling for aspect extraction. arXiv Prepr. arXiv1805.04601.

    Google Scholar 

  70. Huang, B., Y. Ou, and K.M. Carley. 2018. Aspect level sentiment classification with attention-over-attention neural networks. In International Conference on Social Computing, Behavioral-Cultural Modeling and Prediction and Behavior Representation in Modeling and Simulation, 197–206.

    Google Scholar 

  71. Li, X., L. Bing, W. Lam, and B. Shi. 2018. Transformation networks for target-oriented sentiment classification. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 946–956.

    Google Scholar 

  72. Wang, Y., M. Huang, L. Zhao and Others. 2016. Attention-based lstm for aspect-level sentiment classification. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, 606–615.

    Google Scholar 

  73. Ma, D., S. Li, X. Zhang, and H. Wang. 2017. Interactive attention networks for aspect-level sentiment classification. arXiv Prepr. arXiv1709.00893.

    Google Scholar 

  74. Zhang, M., Y. Zhang, and D.-T. Vo. 2016. Gated neural networks for targeted sentiment analysis. In Thirtieth AAAI Conference on Artificial Intelligence.

    Google Scholar 

  75. Mitchell et al. Corpus. http://www.m-mitchell.com/code/index.html.

  76. Xu, J., D. Chen, X. Qiu, and X. Huang. 2016. Cached long short-term memory neural networks for document-level sentiment classification. arXiv Prepr. arXiv1610.04989.

    Google Scholar 

  77. Ma, Y., H. Peng, and E. Cambria. 2018. Targeted aspect-based sentiment analysis via embedding commonsense knowledge into an attentive LSTM. In Thirty-Second AAAI Conference on Artificial Intelligence.

    Google Scholar 

  78. Liu, F., T. Cohn, and T. Baldwin. 2018. Recurrent entity networks with delayed memory update for targeted aspect-based sentiment analysis. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers), 278–283.

    Google Scholar 

  79. Sun, C., L. Huang, and X. Qiu. 2019. Utilizing BERT for aspect-based sentiment analysis via constructing auxiliary sentence. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 380–385.

    Google Scholar 

  80. Pelletier, F.J. 1994. The principle of semantic compositionality. Topoi 13: 11–24.

    Article  MathSciNet  Google Scholar 

  81. Socher, R., and et al. 2013. Recursive deep models for semantic compositionality over a sentiment treebank. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, 1631–1642.

    Google Scholar 

  82. Sentiment Treebank. https://nlp.stanford.edu/sentiment/treebank.html.

  83. Blitzer, J., M. Dredze, and F. Pereira. 2007. Biographies, Bollywood, boom-boxes and blenders: Domain adaptation for sentiment classification. In Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics, 440–447.

    Google Scholar 

  84. Saeidi, M., G. Bouchard, M. Liakata, and S. Riedel. 2016. SentiHood: targeted aspect based sentiment analysis dataset for urban neighbourhoods. In Proceeding COLING 2016, 26th International Conference Computational Linguistics: Technical Papers, 1546–1556.

    Google Scholar 

  85. Crammer, K., and Y. Singer. 2002. Pranking with ranking. In Advances in Neural Information Processing Systems, 641–647.

    Google Scholar 

  86. Moghaddam, S., and M. Ester. 2010. Opinion digger: an unsupervised opinion miner from unstructured product reviews. In Proceedings of the 19th ACM International Conference on Information and Knowledge Management, 1825–1828.

    Google Scholar 

  87. Marcheggiani, D., O. Täckström, A. Esuli, and F. Sebastiani. 2014. Hierarchical multi-label conditional random fields for aspect-oriented opinion mining. In European Conference on Information Retrieval, 273–285.

    Google Scholar 

  88. Lu, B., M. Ott, C. Cardie, and B.K. Tsou. 2011. Multi-aspect sentiment analysis with topic models. In 2011 IEEE 11th International Conference on Data Mining Workshops (ICDMW), 81–88.

    Google Scholar 

  89. Wang, H., Y. Lu, and C. Zhai. 2011. Latent aspect rating analysis without aspect keyword supervision. In Proceedings of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 618–626.

    Google Scholar 

  90. Wang, Q., J. Xu, H. Li, and N. Craswell. 2013. Regularized latent semantic indexing: A new approach to large-scale topic modeling. ACM Transactions on Information Systems (TOIS) 31: 5.

    Article  Google Scholar 

  91. Kullback, S., and R.A. Leibler. 1951. On information and sufficiency. The Annals of Mathematical Statistics 22: 79–86.

    Article  MathSciNet  Google Scholar 

  92. Yelp Dataset. https://www.yelp.com/dataset/challenge.

  93. Diao, Q. and et al. 2014. Jointly modeling aspects, ratings and sentiments for movie recommendation (JMARS). In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 193–202.

    Google Scholar 

  94. Zhang, X., J. Zhao and Y. LeCun. 2015. Character-level convolutional networks for text classification. In Advances in Neural Information Processing Systems, 649–657.

    Google Scholar 

  95. NLP and CC 2013. http://tcci.ccf.org.cn/conference/2013/index.html.

  96. Movie Reviews. http://www.cs.cornell.edu/people/pabo/movie-review-data/.

  97. MPQA Opinion. http://mpqa.cs.pitt.edu.

  98. Go, A., R. Bhayani, and L. Huang. 2009. Twitter sentiment classification using distant supervision. In CS224 N Project Report, Stanford, 1.

    Google Scholar 

  99. Yu, L.-C., and et al. 2016. Building Chinese affective resources in valence-arousal dimensions. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 540–545.

    Google Scholar 

  100. Camera Review. https://www.cs.uic.edu/~liub/FBS/sentiment-analysis.html.

  101. Pontiki, M., and et al. 2016. SemEval-2016 task 5: Aspect based sentiment analysis. In Proceedings of the 10th International Workshop on Semantic Evaluation (SemEval-2016), 19–30.

    Google Scholar 

  102. Dong, L., and et al. 2014. Adaptive recursive neural network for target-dependent twitter sentiment classification. In Proceedings of the 52nd annual meeting of the association for computational linguistics (volume 2: Short papers), 49–54.

    Google Scholar 

  103. Brand24. https://brand24.com.

  104. Clarabridge. https://www.clarabridge.com/platform/analytics/.

  105. Repustate. https://www.repustate.com.

  106. OpenText. https://www.opentext.com/products-and-solutions/products/discovery/information-access-platform/sentiment-analysis.

  107. ParallelDots. https://www.paralleldots.com/sentiment-analysis.

  108. Lexalytics. https://www.lexalytics.com/technology/sentiment-analysis.

  109. Hi-Tech BPO. https://www.hitechbpo.com/sentiment-analysis.php.

  110. Sentiment Analyzer. https://www.danielsoper.com/sentimentanalysis/.

  111. SentiStrength. http://sentistrength.wlv.ac.uk.

  112. Meaning Cloud. https://www.meaningcloud.com/products/sentiment-analysis.

  113. Tweet Sentiment Visualization. https://www.csc2.ncsu.edu/faculty/healey/tweet_viz/tweet_app/.

  114. Rapidminer. https://rapidminer.com/solutions/text-mining/.

  115. Brandwatch. https://www.brandwatch.com/products/analytics/.

  116. Social Mention. http://www.socialmention.com.

  117. Social Searcher. https://www.social-searcher.com/social-buzz/.

  118. Talkwalker’s Quick Search. https://www.talkwalker.com/quick-search-form.

  119. Sentigem. https://sentigem.com.

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Singapore Pte Ltd.

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Pathak, A.R., Agarwal, B., Pandey, M., Rautaray, S. (2020). Application of Deep Learning Approaches for Sentiment Analysis. In: Agarwal, B., Nayak, R., Mittal, N., Patnaik, S. (eds) Deep Learning-Based Approaches for Sentiment Analysis. Algorithms for Intelligent Systems. Springer, Singapore. https://doi.org/10.1007/978-981-15-1216-2_1

Download citation

  • DOI: https://doi.org/10.1007/978-981-15-1216-2_1

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-15-1215-5

  • Online ISBN: 978-981-15-1216-2

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics