SimLex-999 (State of the art)
Jump to navigation
Jump to search
SimLex-999 aims at a cleaner benchmark of similarity (but not relatedness). Pairs of words were chosen to represent different ranges of similarity and with either high or low association. Subjects were instructed to differentiate between similarity and relatedness and rate regarding the former only.
See also: Similarity (State of the art), Similar-Associated-Both Test Collection (State of the art).
Algorithm | Reference for algorithm | Reference for reported results | Type | Spearman's rho | Pearson's r | Notes |
---|---|---|---|---|---|---|
Neural language model | Collobert & Weston (2008)[1] | Hill et al. (2014a)[2] | Distributional | 0.268 | - | Trained on Wikipedia |
Neural language model with global context | Huang et al. (2012)[3] | Hill et al. (2014a)[2] | Distributional | 0.098 | - | Trained on Wikipedia |
Word2vec | Mikolov et al. (2013)[4] | Hill et al. (2014a)[2] | Distributional | 0.414 | - | Trained on Wikipedia |
Lesk | Banjade et al. (2015)[5] | 0.404 | 0.347 | |||
UMBC | Han et al. (2013)[6] | Banjade et al. (2015)[5] | 0.558 | 0.557 | without using POS information | |
SVR4 | Banjade et al. (2015)[5] | Banjade et al. (2015)[5] | Combined | 0.642 | 0.658 | |
ESA | Banjade et al. (2015)[5] | 0.271 | 0.145 | |||
RNNenc | Hill et al. (2014b)[7] | Hill et al. (2014b)[7] | Distributional, multilingual | 0.52 | - |
References
- ↑ R. Collobert and J. Weston. 2008. A unified architecture for natural language pro- cessing: Deep neural networks with multitask learning. In International Conference on Machine Learn- ing, ICML.
- ↑ 2.0 2.1 2.2 Hill, F., Reichart, R., & Korhonen, A. (2014a). SimLex-999: Evaluating Semantic Models with (Genuine) Similarity Estimation. Computation and Language.
- ↑ Eric H Huang, Richard Socher, Christopher D Manning, and Andrew Y Ng. 2012. Improving word representations via global context and multiple word prototypes. In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Long Papers-Volume 1, pages 873–882. Association for Computational Linguistics.
- ↑ Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word representations in vector space. In Proceedings of International Conference of Learning Representations, Scottsdale, Arizona, USA.
- ↑ 5.0 5.1 5.2 5.3 5.4 Banjade, R., Maharjan, N., Niraula, N., Rus, V., & Gautam, D. (2015). Lemon and Tea Are Not Similar: Measuring Word-to-Word Similarity by Combining Different Methods. Computational Linguistics and Intelligent Text Processing, 9041, 335–346. doi:10.1007/978-3-319-18111-0_25
- ↑ Han, L., Kashyap, A., Finin, T., Mayfield, J., Weese, J.: UMBC EBIQUITY-CORE: Semantic textual similarity systems. In: Proceedings of the Second Joint Conference on Lexical and Computational Semantics, vol. 1, pp. 44–52 (2013)
- ↑ 7.0 7.1 Hill, F., Cho, K., Jean, S., Devin, C., & Bengio, Y. (2014b). Not All Neural Embeddings are Born Equal, 1–5.