D16-1175.pdf (232.22 kB)
Improving sparse word representations with distributional inference for semantic composition
chapter
posted on 2023-06-09, 05:14 authored by Thomas Kober, Julie WeedsJulie Weeds, Jeremy ReffinJeremy Reffin, David WeirDavid WeirDistributional models are derived from co- occurrences in a corpus, where only a small proportion of all possible plausible co-occurrences will be observed. This results in a very sparse vector space, requiring a mechanism for inferring missing knowledge. Most methods face this challenge in ways that render the resulting word representations uninterpretable, with the consequence that semantic composition becomes hard to model. In this paper we explore an alternative which involves explicitly inferring unobserved co-occurrences using the distributional neighbourhood. We show that distributional inference improves sparse word repre- sentations on several word similarity benchmarks and demonstrate that our model is competitive with the state-of-the-art for adjective- noun, noun-noun and verb-object compositions while being fully interpretable.
History
Publication status
- Published
File Version
- Published version
Publisher
Association for Computational LinguisticsPublisher URL
Page range
1691-1702Pages
2392.0Event name
Proceedings of the 2016 Conference on Empirical Methods in Natural Language ProcessingEvent location
Austin, TXEvent type
conferenceEvent date
1-5 November 2016Book title
Proceedings of the 2016 Conference on Empirical Methods in Natural Language ProcessingISBN
9781945626258Department affiliated with
- Informatics Publications
Research groups affiliated with
- Data Science Research Group Publications
Full text available
- Yes
Peer reviewed?
- Yes