Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.82
|View full text |Cite
|
Sign up to set email alerts
|

Spelling Error Correction with Soft-Masked BERT

Abstract: Spelling error correction is an important yet challenging task because a satisfactory solution of it essentially needs human-level language understanding ability. Without loss of generality we consider Chinese spelling error correction (CSC) in this paper. A state-ofthe-art method for the task selects a character from a list of candidates for correction (including non-correction) at each position of the sentence on the basis of BERT, the language representation model. The accuracy of the method can be sub-opti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
97
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 130 publications
(104 citation statements)
references
References 14 publications
0
97
0
Order By: Relevance
“…In the experiment, our method outperforms Hong et al (2019) with a large margin, which indicates the effectiveness of the globally optimized chunk-based decoding. Zhang et al (2020) propose to train a detection and a correction network jointly. In the experiment, although they employ 5 million pseudo data for extra pretraining, the proposed method still obtains an improved performance on the correction level.…”
Section: Experiments Results On the Csc Datasetsmentioning
confidence: 99%
See 2 more Smart Citations
“…In the experiment, our method outperforms Hong et al (2019) with a large margin, which indicates the effectiveness of the globally optimized chunk-based decoding. Zhang et al (2020) propose to train a detection and a correction network jointly. In the experiment, although they employ 5 million pseudo data for extra pretraining, the proposed method still obtains an improved performance on the correction level.…”
Section: Experiments Results On the Csc Datasetsmentioning
confidence: 99%
“…generate pseudo data by replacing the character in the training sentence with characters in the confusion set. Similarly, Zhang et al (2020) generate homophonous pseudo data to pretrain the detection and correction network jointly. Web texts are in large quantities and contain more errors than published articles.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Traditional methods of CSC firstly detect misspelled characters and generate candidates via a language model, and then use a phonetic model or rules to filter wrong candidates (Chang, 1995;Chen et al, 2013;Dong et al, 2016). To improve CSC performance, studies mainly focus on two issues: 1) how to improve the language model (Wu et al, 2010;Dong et al, 2016;Zhang et al, 2020) and 2) how to utilize external knowledge of phonological similarity (Jia et al, 2013;Yu and Li, 2014;Cheng et al, 2020). The language model is used to generate fluent sentences and the phonetic features can prevent the model from producing predictions whose pronunciation deviates from that of the original word.…”
Section: Introductionmentioning
confidence: 99%
“…These methods take phonetic information as external knowledge but the discrete candidate selection obstructs the language model from learning directly via backpropagation. Zhang et al (2020) proposed an endto-end CSC model by modifying the mask mechanism of BERT. However, they did not use any phonological information, which is important for exploring words similarity.…”
Section: Introductionmentioning
confidence: 99%