Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.81
|View full text |Cite
|
Sign up to set email alerts
|

SpellGCN: Incorporating Phonological and Visual Similarities into Language Models for Chinese Spelling Check

Abstract: Chinese Spelling Check (CSC) is a task to detect and correct spelling errors in Chinese natural language. Existing methods have made attempts to incorporate the similarity knowledge between Chinese characters. However, they take the similarity knowledge as either an external input resource or just heuristic rules. This paper proposes to incorporate phonological and visual similarity knowledge into language models for CSC via a specialized graph convolutional network (SpellGCN). The model builds a graph over th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
90
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 78 publications
(95 citation statements)
references
References 15 publications
0
90
0
Order By: Relevance
“…In the experiment, although they employ 5 million pseudo data for extra pretraining, the proposed method still obtains an improved performance on the correction level. Cheng et al (2020) propose to incorporate phonological and visual confusion sets into the CSC models through a graph convolutional network. As the performance reported in their paper is obtrained with external training data, we reproduced their results on the standard CSC datasets by rerunning their released code and evaluation scripts.…”
Section: Experiments Results On the Csc Datasetsmentioning
confidence: 99%
See 1 more Smart Citation
“…In the experiment, although they employ 5 million pseudo data for extra pretraining, the proposed method still obtains an improved performance on the correction level. Cheng et al (2020) propose to incorporate phonological and visual confusion sets into the CSC models through a graph convolutional network. As the performance reported in their paper is obtrained with external training data, we reproduced their results on the standard CSC datasets by rerunning their released code and evaluation scripts.…”
Section: Experiments Results On the Csc Datasetsmentioning
confidence: 99%
“…Zhao et al (2017) use conditional random fields (CRFs) to handle two types of misspelled single-character word. Cheng et al (2020) propose to incorporate phonological and visual similarity knowledge into the CSC models via a graph convolutional network.…”
Section: Related Workmentioning
confidence: 99%
“…For the pre-training corpus, we collect a variety of data, such as encyclopedia articles, news, scientific papers, and movie subtitles from a search engine. The CSC training data used in our experiments is the same as Wang et al (2019) and Cheng et al (2020), including three human-annotated training datasets Tseng et al, 2015) and an automatically generated dataset with the approach proposed in Table 1: The performance on SIGHAN13, SIGHAN14, and SIGHAN15 testset. Soft-Masked BERT* is our reproduction of Soft-Masked BERT using the same training data as in our method, while Soft-Masked BERT was trained on an in-house dataset containing 5 million sentences and their counterparts with automatically generated errors, as reported in Zhang et al (2020), where the authors only provided their results on SIGHAN15.…”
Section: Data Processingmentioning
confidence: 99%
“…• SpellGCN (Cheng et al, 2020) incorporates two similarity graphs into a pre-trained sequence-labeling model via graph convolutional network. The two graphs are derived from a confusion set and correspond to pronunciation and shape similarities.…”
Section: Model Settingsmentioning
confidence: 99%
See 1 more Smart Citation