Abstract
Embedding algorithms work by optimizing the distance between a word and its context(s), generating an embedding space that encodes their distributional representation. In addition to single words or word pieces, other features, which result from a deeper analysis of the text, can be used to enrich such representations with additional information. Such features are influenced by the tokenization strategy used to chunk the text and can include not only lexical and part-of-speech information but also annotations about the disambiguated sense of a word according to a structured knowledge graph. In this chapter we analyze the impact that explicitly adding lexical, grammatical and semantic information during the training of Vecsigrafo has in the resulting representations and whether or not this can enhance their downstream performance. To illustrate this analysis we focus on corpora from the scientific domain, where rich, multi-word expressions are frequent, hence requiring advanced tokenization strategies.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Gomez-Perez, J.M., Denaux, R., Garcia-Silva, A. (2020). Capturing Lexical, Grammatical, and Semantic Information with Vecsigrafo. In: A Practical Guide to Hybrid Natural Language Processing. Springer, Cham. https://doi.org/10.1007/978-3-030-44830-1_8
Download citation
DOI: https://doi.org/10.1007/978-3-030-44830-1_8
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-44829-5
Online ISBN: 978-3-030-44830-1
eBook Packages: Computer ScienceComputer Science (R0)