Proceedings of the 5th Workshop on Natural Language Processing Techniques for Educational Applications 2018
DOI: 10.18653/v1/w18-3707
|View full text |Cite
|
Sign up to set email alerts
|

Chinese Grammatical Error Diagnosis using Statistical and Prior Knowledge driven Features with Probabilistic Ensemble Enhancement

Abstract: This paper describes our system at NLPTEA-2018 Task #1: Chinese Grammatical Error Diagnosis. Grammatical Error Diagnosis is one of the most challenging NLP tasks,which is to locate grammar errors and tell error types. Our system is built on the model of bidirectional Long Short-Term Memory with a conditional random field layer (BiLSTM-CRF) but integrates with several new features. First, richer features are considered in the BiLSTM-CRF model; second, a probabilistic ensemble approach is adopted; third, Templat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
6
2
1

Relationship

1
8

Authors

Journals

citations
Cited by 13 publications
(6 citation statements)
references
References 5 publications
0
5
0
Order By: Relevance
“…Most studies regard it as a sequence tagging task, where each token will be given a correct label or an error-type. Sequence labeling methods are widely used for CGED, such as feature-based statistical models (Chang et al, 2012), and neural models (Fu et al, 2018). Due to the effectiveness of BERT (Devlin et al, 2019) in many other NLP applications, recent studies adopt BERT as the basic architecture of CGED models (Fang et al, 2020;Wang et al, 2020b;Li and Shi, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…Most studies regard it as a sequence tagging task, where each token will be given a correct label or an error-type. Sequence labeling methods are widely used for CGED, such as feature-based statistical models (Chang et al, 2012), and neural models (Fu et al, 2018). Due to the effectiveness of BERT (Devlin et al, 2019) in many other NLP applications, recent studies adopt BERT as the basic architecture of CGED models (Fang et al, 2020;Wang et al, 2020b;Li and Shi, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…Many researchers have made outstanding achievements on CSC (Zhang et al, 2020; and CGED (Fu et al, 2018). Existing CSC and CGED models cannot achieve good results for CSER because semantic errors are often difficult compared to other errors.…”
Section: Text Error Detectionmentioning
confidence: 99%
“…Previous research [2,5] put a lot of effort into feature engineering, such as pretrained and parsing features. The most significant parsing characteristics are part-of-speech tagging (POS) and dependency information, indicating that the job is strongly related to the structure of the sentence syntactic dependence.…”
Section: Gcnmentioning
confidence: 99%