“…We denote them as GECToR-BERT, GECToR-XLNet and GECToR (Ensemble) respectively. For the Chinese GEC task, we compare S2A to several best performing systems evaluated on the NLPCC-2018 dataset, including three top systems in the NLPCC-2018 challenge (YouDao (Fu, Huang, and Duan 2018), AliGM (Zhou et al 2018), BLCU (Ren, Yang, and Xun 2018)), the seq2seq baseline Char Transformer, and the current state-of-the-art method MaskGEC (Zhao and Wang 2020). Note that the proposed S2A model is orthogonal to MaskGEC, and we also report our results enhanced with the data augmentation method of MaskGEC.…”