Proceedings of the Joint Workshop on Multiword Expressions and WordNet (MWE-WN 2019) 2019
DOI: 10.18653/v1/w19-5111
|View full text |Cite
|
Sign up to set email alerts
|

A Systematic Comparison of English Noun Compound Representations

Abstract: Building meaningful representations of noun compounds is not trivial since many of them scarcely appear in the corpus. To that end, composition functions approximate the distributional representation of a noun compound by combining its constituent distributional vectors. In the more general case, phrase embeddings have been trained by minimizing the distance between the vectors representing paraphrases. We compare various types of noun compound representations, including distributional, compositional, and para… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
13
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(13 citation statements)
references
References 20 publications
0
13
0
Order By: Relevance
“…Much like distributional similarity approaches for learning word representations (Mikolov et al, 2013a;Bojanowski et al, 2017;Pennington et al, 2014, inter alia), a semantic representation of MWEs can be trained using a distributional approach that treats MWEs as single tokens (Mikolov et al, 2013b). However, this approach cannot handle out of vocabulary (OOV) MWEs, and it is likely to suffer from sparsity (Shwartz, 2019), particularly as the MWEs grow in length.…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…Much like distributional similarity approaches for learning word representations (Mikolov et al, 2013a;Bojanowski et al, 2017;Pennington et al, 2014, inter alia), a semantic representation of MWEs can be trained using a distributional approach that treats MWEs as single tokens (Mikolov et al, 2013b). However, this approach cannot handle out of vocabulary (OOV) MWEs, and it is likely to suffer from sparsity (Shwartz, 2019), particularly as the MWEs grow in length.…”
Section: Related Workmentioning
confidence: 99%
“…Our proposed approach does not suffer from this observer effect, as we learn our compositional function indirectly (through Skip-Gram), without relying on a distributionally learned embedding for MWEs for training. Shwartz (2019) also avoid this reliance on the gold embedding of the multi-word, learning the function indirectly. The compositional function is used to encode the multiword and its paraphrase and is then trained to maximize the cosine similarity between the encodings.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations