Neural Latent Relational Analysis to Capture Lexical Semantic Relations in a Vector Space

Koki Washio, Tsuneaki Kato


Abstract
Capturing the semantic relations of words in a vector space contributes to many natural language processing tasks. One promising approach exploits lexico-syntactic patterns as features of word pairs. In this paper, we propose a novel model of this pattern-based approach, neural latent relational analysis (NLRA). NLRA can generalize co-occurrences of word pairs and lexico-syntactic patterns, and obtain embeddings of the word pairs that do not co-occur. This overcomes the critical data sparseness problem encountered in previous pattern-based models. Our experimental results on measuring relational similarity demonstrate that NLRA outperforms the previous pattern-based models. In addition, when combined with a vector offset model, NLRA achieves a performance comparable to that of the state-of-the-art model that exploits additional semantic relational data.
Anthology ID:
D18-1058
Volume:
Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing
Month:
October-November
Year:
2018
Address:
Brussels, Belgium
Editors:
Ellen Riloff, David Chiang, Julia Hockenmaier, Jun’ichi Tsujii
Venue:
EMNLP
SIG:
SIGDAT
Publisher:
Association for Computational Linguistics
Note:
Pages:
594–600
Language:
URL:
https://aclanthology.org/D18-1058
DOI:
10.18653/v1/D18-1058
Bibkey:
Cite (ACL):
Koki Washio and Tsuneaki Kato. 2018. Neural Latent Relational Analysis to Capture Lexical Semantic Relations in a Vector Space. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 594–600, Brussels, Belgium. Association for Computational Linguistics.
Cite (Informal):
Neural Latent Relational Analysis to Capture Lexical Semantic Relations in a Vector Space (Washio & Kato, EMNLP 2018)
Copy Citation:
PDF:
https://aclanthology.org/D18-1058.pdf
Video:
 https://aclanthology.org/D18-1058.mp4