2020
DOI: 10.48550/arxiv.2005.07421
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Spelling Error Correction with Soft-Masked BERT

Abstract: Spelling error correction is an important yet challenging task because a satisfactory solution of it essentially needs human-level language understanding ability. Without loss of generality we consider Chinese spelling error correction (CSC) in this paper. A state-ofthe-art method for the task selects a character from a list of candidates for correction (including non-correction) at each position of the sentence on the basis of BERT, the language representation model. The accuracy of the method can be sub-opti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
20
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 11 publications
(20 citation statements)
references
References 16 publications
0
20
0
Order By: Relevance
“…In the end, a contextsensitive model is used to score all candidates and pick up the best one. In [12], Bi-GRU network is used to assign error occurrence probability for each input token. Then vector representation of each token is interpolated with a special mask token representation according to those calculated probabilities.…”
Section: Pipeline Text-based Correction Methodsmentioning
confidence: 99%
“…In the end, a contextsensitive model is used to score all candidates and pick up the best one. In [12], Bi-GRU network is used to assign error occurrence probability for each input token. Then vector representation of each token is interpolated with a special mask token representation according to those calculated probabilities.…”
Section: Pipeline Text-based Correction Methodsmentioning
confidence: 99%
“…CSC Data Augmentation: In order to make up for the lack of labeled data, previous studies usually build additional pseudo data to increase the performance. The mainstream method is based on the confusion set Zhang et al, , 2020, the pseudo data constructed in this way is extensive in size but relatively low in quality because of the big gap from the true error distribution. Another relatively high-quality construction method is based on ASR or OCR method (Wang et al, 2018).…”
Section: Related Workmentioning
confidence: 99%
“…Various GEC studies modified and utilized the mask mechanism of BERT [7] to detect the errors and further correct them [8,9]. Asano [35] incorporated the BERT to detect sentences with grammatical errors.…”
Section: Related Workmentioning
confidence: 99%
“…Prior GEC and GED studies have obtained the outstanding achievements in this area. Most of them employed n-gram [2,3], confusion set [4,5], language model [6] including the BERT [7][8][9] etc. to diagnose the errors.…”
Section: Introductionmentioning
confidence: 99%