Abstract
Neural word embeddings have been widely used in modern NLP applications as they provide vector representation of words and capture the semantic properties of words and the linguistic relationship between the words. Many research groups have released their own version of word embeddings. However, they are trained on generic corpora, which limits their direct use for domain specific tasks. In this paper, we evaluate a set of pretrained word embeddings which were provided to us, on a standard NLP task - Sentiment Polarity Identification Task.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Almuhareb, A.: Attributes in lexical acquisition. Ph.D. thesis, University of Essex (2006)
Baroni, M., Dinu, G., Kruszewski, G.: Don’t count, predict! a systematic comparison of context-counting vs. context-predicting semantic vectors. In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), vol. 1, pp. 238–247 (2014)
Cinková, Silvie: WordSim353 for czech. In: Sojka, Petr, Horák, Aleš, Kopeček, Ivan, Pala, Karel (eds.) TSD 2016. LNCS (LNAI), vol. 9924, pp. 190–197. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-45510-5_22
Collobert, R., Weston, J., Bottou, L., Karlen, M., Kavukcuoglu, K., Kuksa, P.: Natural language processing (almost) from scratch. J. Mach. Learn. Res. 12(Aug), 2493–2537 (2011)
Dhillon, P., Rodu, J., Foster, D., Ungar, L.: Two step CCA: a new spectral method for estimating vector models of words. arXiv preprint arXiv:1206.6403 (2012)
Ghannay, S., Favre, B., Esteve, Y., Camelin, N.: Word embedding evaluation and combination
Lample, G., Ballesteros, M., Subramanian, S., Kawakami, K., Dyer, C.: Neural architectures for named entity recognition. arXiv preprint arXiv:1603.01360 (2016)
Landauer, T.K., Dumais, S.T.: A solution to plato’s problem: the latent semantic analysis theory of acquisition, induction, and representation of knowledge. Psychol. Rev. 104(2), 211 (1997)
Lebret, R., Collobert, R.: Word emdeddings through hellinger PCA. arXiv preprint arXiv:1312.5542 (2013)
Levy, O., Goldberg, Y.: Dependency-based word embeddings. In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), vol. 2, pp. 302–308 (2014)
Li, P., Hastie, T.J., Church, K.W.: Very sparse random projections. In: Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 287–296. ACM (2006)
Maas, A.L., Daly, R.E., Pham, P.T., Huang, D., Ng, A.Y., Potts, C.: Learning word vectors for sentiment analysis. In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, vol. 1, pp. 142–150. Association for Computational Linguistics (2011)
Mikolov, T., Yih, W.t., Zweig, G.: Linguistic regularities in continuous space word representations. In: Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 746–751 (2013)
Nayak, N., Angeli, G., Manning, C.D.: Evaluating word embeddings using a representative suite of practical tasks. In: Proceedings of the 1st Workshop on Evaluating Vector-Space Representations for NLP, pp. 19–23 (2016)
Padó, S., Lapata, M.: Dependency-based construction of semantic space models. Comput. Linguist. 33(2), 161–199 (2007)
Pennington, J., Socher, R., Manning, C.: Glove: Global vectors for word representation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1532–1543 (2014)
Ren, M., Kiros, R., Zemel, R.: Exploring models and data for image question answering. In: Advances in Neural Information Processing Systems, pp. 2953–2961 (2015)
Rubenstein, H., Goodenough, J.B.: Contextual correlates of synonymy. Commun. ACM 8(10), 627–633 (1965)
Santos, C.N.d., Guimaraes, V.: Boosting named entity recognition with neural character embeddings. arXiv preprint arXiv:1505.05008 (2015)
Schnabel, T., Labutov, I., Mimno, D., Joachims, T.: Evaluation methods for unsupervised word embeddings. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 298–307 (2015)
Schwenk, H.: CSLM-a modular open-source continuous space language modeling toolkit (2013)
Tang, D., Wei, F., Yang, N., Zhou, M., Liu, T., Qin, B.: Learning sentiment-specific word embedding for twitter sentiment classification. In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), vol. 1, pp. 1555–1565 (2014)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Indurthi, V., Oota, S.R. (2018). Evaluating Quality of Word Embeddings with Sentiment Polarity Identification Task. In: Buscaldi, D., Gangemi, A., Reforgiato Recupero, D. (eds) Semantic Web Challenges. SemWebEval 2018. Communications in Computer and Information Science, vol 927. Springer, Cham. https://doi.org/10.1007/978-3-030-00072-1_18
Download citation
DOI: https://doi.org/10.1007/978-3-030-00072-1_18
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-00071-4
Online ISBN: 978-3-030-00072-1
eBook Packages: Computer ScienceComputer Science (R0)