Abstract
Distributed word representations in the form of dense vectors, known as word embeddings, are the basic building blocks for machine-learning based natural language processing. Such embeddings play an important role in tasks such as part-of-speech tagging, chunking, named entity recognition, and semantic role labeling, as well as downstream tasks including sentiment analysis and more in general text classification. However, early word embeddings were static context-independent representations that fail to capture multiple meanings for polysemous words. This chapter presents an overview of such traditional word embeddings, but also of alternative approaches that have been proposed to produce sense and concept embeddings using disambiguated corpora or directly from knowledge graphs. As a result, this chapter serves as a conceptual framework for the rest of book.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Gomez-Perez, J.M., Denaux, R., Garcia-Silva, A. (2020). Word, Sense, and Graph Embeddings. In: A Practical Guide to Hybrid Natural Language Processing. Springer, Cham. https://doi.org/10.1007/978-3-030-44830-1_2
Download citation
DOI: https://doi.org/10.1007/978-3-030-44830-1_2
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-44829-5
Online ISBN: 978-3-030-44830-1
eBook Packages: Computer ScienceComputer Science (R0)