Abstract
In previous chapters we have seen a variety of ways to train models to derive embedding spaces for words and concepts and other nodes in knowledge graphs. As you often do not have control over the full training procedure, you may find yourself with several embedding spaces which have (conceptually) overlapping vocabularies. How can you best combine such embedding spaces?. In this chapter we look at various techniques for aligning disparate embedding spaces. This is particularly useful in hybrid settings like when using embedding spaces for knowledge graph curation and interlinking.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Gomez-Perez, J.M., Denaux, R., Garcia-Silva, A. (2020). Aligning Embedding Spaces and Applications for Knowledge Graphs. In: A Practical Guide to Hybrid Natural Language Processing. Springer, Cham. https://doi.org/10.1007/978-3-030-44830-1_9
Download citation
DOI: https://doi.org/10.1007/978-3-030-44830-1_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-44829-5
Online ISBN: 978-3-030-44830-1
eBook Packages: Computer ScienceComputer Science (R0)