Proceedings of the 2018 Conference of the North American Chapter Of the Association for Computational Linguistics: Hu 2018
DOI: 10.18653/v1/n18-2029
|View full text |Cite
|
Sign up to set email alerts
|

Discriminating between Lexico-Semantic Relations with the Specialization Tensor Model

Abstract: We present a simple and effective feedforward neural architecture for discriminating between lexico-semantic relations (synonymy, antonymy, hypernymy, and meronymy). Our Specialization Tensor Model (STM) simultaneously produces multiple different specializations of input distributional word vectors, tailored for predicting lexico-semantic relations for word pairs. STM outperforms more complex state-of-the-art architectures on two benchmark datasets and exhibits stable performance across languages. We also show… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
27
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 21 publications
(27 citation statements)
references
References 19 publications
(19 reference statements)
0
27
0
Order By: Relevance
“…The 42% remaining cases involve a number of challenging antonyms and synonyms, but also a few remarkably easy cases. The latter include mor-phologically marked antonyms (e.g., Richtigkeit 'correctness' vs. Unrichtigkeit 'incorrectness') and instances of synonyms that involve two identical words (e.g., Verwandtschaft/Verwandtschaft 'kinship'), which presumably is an artifact of the automatic translation step used in creating the German dataset (see Glavaš and Vulić (2018)). Both types of errors could be identified by simple heuristics but are incorrectly classified by the model.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…The 42% remaining cases involve a number of challenging antonyms and synonyms, but also a few remarkably easy cases. The latter include mor-phologically marked antonyms (e.g., Richtigkeit 'correctness' vs. Unrichtigkeit 'incorrectness') and instances of synonyms that involve two identical words (e.g., Verwandtschaft/Verwandtschaft 'kinship'), which presumably is an artifact of the automatic translation step used in creating the German dataset (see Glavaš and Vulić (2018)). Both types of errors could be identified by simple heuristics but are incorrectly classified by the model.…”
Section: Discussionmentioning
confidence: 99%
“…Supervision for lexical classification tasks is not available in all languages. To overcome this difficulty, some approaches (Mrkšić et al, 2017;Glavaš and Vulić, 2018) combine resources for English and cross-lingual word embeddings to distinguish lexical relations in other languages. That is, they train and test one model across different languages.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…There are broadly two strategies to solve this problem. One line of work (Glavaš and Vulić, 2018;Mrkšić et al, 2017) uses model transfer, where a single model is trained on data from a high-resource language, and is then ported to the target language using cross-lingual embeddings. In contrast, Roth and Upadhyay (2019) translate training data from English into a target language using a combination of unsupervised cross-lingual embeddings (Artetxe et al, 2018) and monolingual information from the target language.…”
Section: Related Workmentioning
confidence: 99%
“…Apart from Hearst patterns (Hearst, 1992), there are few high-quality textual patterns to recognize lexical relations other than hypernymy. Distributional approaches consider the global contexts of terms to predict lexical relations using word embeddings (Baroni et al, 2012;Glavas and Vulic, 2018). They are reported to outperform several path-based approaches, but can suffer from the "lexical memorization" problem (Levy et al, 2015;.…”
Section: Introductionmentioning
confidence: 99%