2017
DOI: 10.1162/tacl_a_00063
|View full text |Cite
|
Sign up to set email alerts
|

Semantic Specialization of Distributional Word Vector Spaces using Monolingual and Cross-Lingual Constraints

Abstract: We present ATTRACT-REPEL, an algorithm for improving the semantic quality of word vectors by injecting constraints extracted from lexical resources. ATTRACT-REPEL facilitates the use of constraints from mono-and crosslingual resources, yielding semantically specialised cross-lingual vector spaces. Our evaluation shows that the method can make use of existing cross-lingual lexicons to construct highquality vector spaces for a plethora of different languages, facilitating semantic transfer from high-to lower-res… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

4
185
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 161 publications
(189 citation statements)
references
References 48 publications
(87 reference statements)
4
185
0
Order By: Relevance
“…Further, a lot of syntactic and semantic parsing models recently successfully incorporated parameter sharing for training parsers in closely related languges (Duong et al, 2015;Ammar et al, 2016;Susanto and Lu, 2017;. In the domain of dialog managers, Mrkšić et al (2017) and Chen et al (2018) presented methods for crosslingual transfer for dialog state tracking.…”
Section: Related Workmentioning
confidence: 99%
“…Further, a lot of syntactic and semantic parsing models recently successfully incorporated parameter sharing for training parsers in closely related languges (Duong et al, 2015;Ammar et al, 2016;Susanto and Lu, 2017;. In the domain of dialog managers, Mrkšić et al (2017) and Chen et al (2018) presented methods for crosslingual transfer for dialog state tracking.…”
Section: Related Workmentioning
confidence: 99%
“…Also note that unlike retro-fitting and similar techniques (Rothe and Schütze, 2015;Pilehvar and Collier, 2016;Mrkšić et al, 2017), our approach does not use any training corpus or pretrained input embeddings. The synset representations are trained on the WordNet graph alone.…”
Section: Related Workmentioning
confidence: 99%
“…WOZ2.0 consists of a total of 1200 dialogues, out of which 600 are for training, 200 for development and 400 for testing. [17] translated the WOZ.0 English data both to German and Italian using professional translators. We experiment on the three languages (English, German, Italian) of the WOZ2.0 dataset.…”
Section: Datasetsmentioning
confidence: 99%