Proceedings of the 55th Annual Meeting of the Association For Computational Linguistics (Volume 2: Short Papers) 2017
DOI: 10.18653/v1/p17-2096
|View full text |Cite
|
Sign up to set email alerts
|

Fast and Accurate Neural Word Segmentation for Chinese

Abstract: Neural models with minimal feature engineering have achieved competitive performance against traditional methods for the task of Chinese word segmentation. However, both training and working procedures of the current neural models are computationally inefficient. This paper presents a greedy neural word segmenter with balanced word and character embedding inputs to alleviate the existing drawbacks. Our segmenter is truly end-toend, capable of performing segmentation much faster and even more accurate than stat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
92
0
1

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
3
1

Relationship

5
4

Authors

Journals

citations
Cited by 94 publications
(93 citation statements)
references
References 27 publications
0
92
0
1
Order By: Relevance
“…For example, Liu et al (2016) runs bi-directional LSTM over characters of the word candidate and then concatenate bi-directional LSTM outputs at both end points. Cai et al (2017) adopts a gating mechanism to control relative importance of each character in the word candidate. Besides modeling word representation directly, sequential labeling is another popular approach.…”
Section: Modelmentioning
confidence: 99%
See 1 more Smart Citation
“…For example, Liu et al (2016) runs bi-directional LSTM over characters of the word candidate and then concatenate bi-directional LSTM outputs at both end points. Cai et al (2017) adopts a gating mechanism to control relative importance of each character in the word candidate. Besides modeling word representation directly, sequential labeling is another popular approach.…”
Section: Modelmentioning
confidence: 99%
“…Neural networks have become ubiquitous in natural language processing. For the word segmentation task, there has been a growing body of work exploring novel neural network architectures for learning useful representation and thus better segmentation prediction (Pei et al, 2014;Ma and Hinrichs, 2015;Zhang et al, 2016a;Liu et al, 2016;Cai et al, 2017;Wang and Xu, 2017).…”
Section: Introductionmentioning
confidence: 99%
“…5 CMRC-2017 Leaderboard: http://www.hfl-tek.com/cmrc2017/ leaderboard/. 6 The word vocabulary sizes of SNLI and CMRC-2017 are 30k and 90k respectively. Table VI, which shows our Word + BPE-FRQ significantly outperforms the CAS Reader in all types of testing, with improvements of 7.0% on PD and 8.8% on CFT test sets, respectively.…”
Section: B Reading Comprehensionmentioning
confidence: 99%
“…To alleviate the noise issue introduced by the extra part in the source side, inspired by the work of (Dhingra et al, 2016;Pang et al, 2016;Zhang et al, 2018c,a,b;Cai et al, 2017b), our model adopts a gated-attention (GA) mechanism that performs multiple hops over the pinyin with the extended context as shown in Figure 1(d).…”
Section: Modelmentioning
confidence: 99%