Proceedings of the Joint Workshop on Multiword Expressions and WordNet (MWE-WN 2019) 2019
DOI: 10.18653/v1/w19-5121
|View full text |Cite
|
Sign up to set email alerts
|

The Impact of Word Representations on Sequential Neural MWE Identification

Abstract: Recent initiatives such as the PARSEME shared task have allowed the rapid development of MWE identification systems. Many of those are based on recent NLP advances, using neural sequence models that take continuous word representations as input. We study two related questions in neural verbal MWE identification: (a) the use of lemmas and/or surface forms as input features, and (b) the use of word-based or character-based embeddings to represent them. Our experiments on Basque, French, and Polish show that char… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
4
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(8 citation statements)
references
References 16 publications
(17 reference statements)
0
4
0
Order By: Relevance
“…Figure 1 shows the distribution of papers across the 24 languages considered by our paper sample. The reasons that lead to choosing a given corpus and/or set of languages in non-ST works are various: language diversity (Zampieri et al, 2019), corpus domain (Liu et al, 2021), and corpus quality and size (Pasquer et al, 2020b).…”
Section: Corpus Constitution and Selectionmentioning
confidence: 99%
See 3 more Smart Citations
“…Figure 1 shows the distribution of papers across the 24 languages considered by our paper sample. The reasons that lead to choosing a given corpus and/or set of languages in non-ST works are various: language diversity (Zampieri et al, 2019), corpus domain (Liu et al, 2021), and corpus quality and size (Pasquer et al, 2020b).…”
Section: Corpus Constitution and Selectionmentioning
confidence: 99%
“…6 Most papers do not explicitly mention their strategy to deal with overlapping MWEs. When mentioned, overlapping MWE annotations are either ignored (Zampieri et al, 2022a), duplicated into separate sentences (Zampieri et al, 2018), or handled by the tagging scheme (Yirmibeşoglu and Güngör, 2020).…”
Section: Corpus and Splitsmentioning
confidence: 99%
See 2 more Smart Citations
“…Figure 1 shows the distribution of papers across the 24 languages considered by our paper sample. The reasons that lead to choosing a given corpus and/or set of languages in non-ST works are various: language diversity (Zampieri et al, 2019), corpus domain , and corpus quality and size (Pasquer et al, 2020b).…”
Section: Corpus Constitution and Selectionmentioning
confidence: 99%