Proceedings of the 17th International Conference on Parsing Technologies and the IWPT 2021 Shared Task on Parsing Into Enhanced 2021
DOI: 10.18653/v1/2021.iwpt-1.13
|View full text |Cite
|
Sign up to set email alerts
|

Applying Occam’s Razor to Transformer-Based Dependency Parsing: What Works, What Doesn’t, and What is Really Necessary

Abstract: The introduction of pre-trained transformerbased contextualized word embeddings has led to considerable improvements in the accuracy of graph-based parsers for frameworks such as Universal Dependencies (UD). However, previous works differ in various dimensions, including their choice of pre-trained language models and whether they use LSTM layers. With the aims of disentangling the effects of these choices and identifying a simple yet widely applicable architecture, we introduce STEPS, a new modular graph-base… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
1
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 36 publications
0
2
0
Order By: Relevance
“…ILSP Neural NLP Toolkit for Greek (Prokopidis & Piperidis, 2020) HR (Ljubešić & Dobrovoljc, 2019;Terčon & Ljubešić, 2023) HU ▲ •• huspacy (Orosz, Szántó, Berkecz, Szabó, & Farkas, 2022) IS Stanza (Qi et al, 2020) LV (Ljubešić & Dobrovoljc, 2019;Terčon & Ljubešić, 2023) TR Çöltekin, 2010), ▲ steps-parser (Grünewald, Friedrich, & Kuhn, 2021), TurkishNER UA…”
Section: Compiling Individual Corporamentioning
confidence: 99%
“…ILSP Neural NLP Toolkit for Greek (Prokopidis & Piperidis, 2020) HR (Ljubešić & Dobrovoljc, 2019;Terčon & Ljubešić, 2023) HU ▲ •• huspacy (Orosz, Szántó, Berkecz, Szabó, & Farkas, 2022) IS Stanza (Qi et al, 2020) LV (Ljubešić & Dobrovoljc, 2019;Terčon & Ljubešić, 2023) TR Çöltekin, 2010), ▲ steps-parser (Grünewald, Friedrich, & Kuhn, 2021), TurkishNER UA…”
Section: Compiling Individual Corporamentioning
confidence: 99%
“…So, we reimplement the auxiliary task modules and the combined parsing approach for an XLM-R-based encoding module. For this purpose we follow the XLM-R-based parsing architecture of Grünewald et al (2021) which has the same biaffine parsing model described in Dozat and Manning (2017). Our aim is to observe how extracting parsingrelated knowledge from semi-supervised auxiliary tasks affects a multilingual transformer model.…”
Section: Transformer-based Adaptation Of the Modelmentioning
confidence: 99%