Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2023
DOI: 10.18653/v1/2023.acl-long.86
|View full text |Cite
|
Sign up to set email alerts
|

GEC-DePenD: Non-Autoregressive Grammatical Error Correction with Decoupled Permutation and Decoding

Abstract: Grammatical error correction (GEC) is an important NLP task that is currently usually solved with autoregressive sequence-tosequence models. However, approaches of this class are inherently slow due to one-byone token generation, so non-autoregressive alternatives are needed. In this work, we propose a novel non-autoregressive approach to GEC that decouples the architecture into a permutation network that outputs a self-attention weight matrix that can be used in beam search to find the best permutation of inp… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
1
1
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 27 publications
0
4
0
Order By: Relevance
“…Recent advances in GEC also explore non-autoregressive models, which provide faster performance by decoupling the error detection and correction processes, allowing for more dynamic responses to the identified errors [11]. The essence of this task can be succinctly summarized through the following formulation: Given a sentence S err laden with grammatical inaccuracies, the objective of English GEC is to generate a grammatically correct sentence S corr , wherein S corr retains the original intent and content of S err to the greatest extent possible.…”
Section: Gecmentioning
confidence: 99%
“…Recent advances in GEC also explore non-autoregressive models, which provide faster performance by decoupling the error detection and correction processes, allowing for more dynamic responses to the identified errors [11]. The essence of this task can be succinctly summarized through the following formulation: Given a sentence S err laden with grammatical inaccuracies, the objective of English GEC is to generate a grammatically correct sentence S corr , wherein S corr retains the original intent and content of S err to the greatest extent possible.…”
Section: Gecmentioning
confidence: 99%
“…Gu et al (2019) non-autoregressively refine an output sequence using language-agnostic insertions and deletions. Yakovlev et al (2023) decompose the inference stage into permutation and decoding. First, a permutation network repositions the tokens of an input sequence with possible deletions and inser-tions.…”
Section: Synthetic Datamentioning
confidence: 99%
“…Prior research incorporates detection results as supplementary information for Seq2Seq correction models Yuan et al, 2021b;Li et al, 2023a). The methods proposed by Mallinson et al (2020) and Yakovlev et al (2023) employ the Masked Language Model (MLM) (Devlin et al, 2018) to obtain corrections, which are constrained by mask number. Chen et al (2020) introduces error span detection and correction to address the GEC problem, which allows for flexible corrections while maximizing time efficiency.…”
Section: Detection-correction Gecmentioning
confidence: 99%
“…Multi-Encoder (Yuan et al, 2021a) encodes error categories as auxiliary information. GEC-DePend (Yakovlev et al, 2023) with correction by the MLM. TemplateGEC (Li et al, 2023a) uses the output of the GECToR model as supplementary information for Seq2Seq models.…”
Section: Model Settingsmentioning
confidence: 99%