2020
DOI: 10.1016/j.patrec.2020.07.017
|View full text |Cite
|
Sign up to set email alerts
|

Encoding multi-granularity structural information for joint Chinese word segmentation and POS tagging

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
13
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 16 publications
(13 citation statements)
references
References 3 publications
0
13
0
Order By: Relevance
“…The former is mainly based on word‐based rule matching for a given constructed dictionary, such as positive maximum matching rules, reverse maximum matching rules (Luo et al., 2018) and bidirectional matching rules (Huang et al., 2015; Yunita et al., 2010). The latter is trained on annotated Chinese text to obtain different models: Hidden Markov models (HMMs) and Conditional Random Fields (CRFs), statistical machine learning models (Du et al., 2018; Huang et al., 2017; Liang et al., 2019; Y. Liu et al., 2014; Zhang & Li, 2016), deep learning models (Xu & Sun, 2016; Zhao et al., 2020), etc. Based on the trained model, the text of the unknown label is segmented.…”
Section: Related Workmentioning
confidence: 99%
“…The former is mainly based on word‐based rule matching for a given constructed dictionary, such as positive maximum matching rules, reverse maximum matching rules (Luo et al., 2018) and bidirectional matching rules (Huang et al., 2015; Yunita et al., 2010). The latter is trained on annotated Chinese text to obtain different models: Hidden Markov models (HMMs) and Conditional Random Fields (CRFs), statistical machine learning models (Du et al., 2018; Huang et al., 2017; Liang et al., 2019; Y. Liu et al., 2014; Zhang & Li, 2016), deep learning models (Xu & Sun, 2016; Zhao et al., 2020), etc. Based on the trained model, the text of the unknown label is segmented.…”
Section: Related Workmentioning
confidence: 99%
“…Shao proposed a bidirectional RNN-CRF architecture that incorporated rich contextual information and sub-character level features [8]. Zhao presented a model based on lattice-LSTM and Convolutional Network, exploiting character, word, and subword information [9]. Tian proposed a two-way attention neural network model using context features and their corresponding syntactic information of characters in the sequence [10].…”
Section: Related Workmentioning
confidence: 99%
“…It is a character-based model and utilizes rich contextual information and sub-character level features [8]. Zhao presented a lattice-LSTM and Convolutional Network, which can exploit multi-granularity of information, including characters, words, and subwords [9]. Tian introduced a neural network model with a two-way attention mechanism.…”
Section: Introductionmentioning
confidence: 99%
“…Park et al [15] proposed regularization of subwords, using a unary language model to generate multiple candidate subword sequences, enriching the input of the encoder to enhance the robustness of the translation system. Zhao et al [16] introduced the representation of multigranularity BPE to obtain the semantic representation of vocabulary on average. Zhang et al [17] believed that the encoder word vector layer, decoder word vector layer, and decoder output layer have different functions, so the choice of BPE granularity for different layers should also be different.…”
Section: Introductionmentioning
confidence: 99%