2020
DOI: 10.1016/j.neucom.2020.07.112
|View full text |Cite
|
Sign up to set email alerts
|

A dual channel class hierarchy based recurrent language modeling

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 27 publications
0
2
0
Order By: Relevance
“…In the case of English, since the basic unit of English is the word, it is only necessary to split the word directly according to the space. However, since English sentences contain stop words, they also need to be deactivated during the word separation process [29].…”
Section: Split Word Processing As (Iwslt) 2019 Dataset Contains Chine...mentioning
confidence: 99%
“…In the case of English, since the basic unit of English is the word, it is only necessary to split the word directly according to the space. However, since English sentences contain stop words, they also need to be deactivated during the word separation process [29].…”
Section: Split Word Processing As (Iwslt) 2019 Dataset Contains Chine...mentioning
confidence: 99%
“…Transformers have outperformed most of the Natural Language Processing tasks. However, language modelling requires Model Validation Test RNN-LDA + KN-5 + cache [44] -92.0 LSTM (large) [45] 82.2 78.4 Variational LSTM (large, MC) [46] -73.4 CharCNN [47] -78.9 Variational LSTM (tied) + augmented loss [48] 71.1 68.5 Variational RHN (tied) [49] 67.9 65.4 NAS Cell (tied) [50] -62.4 4-layer skip connection LSTM (tied) [51] 60.9 58.3 AWD-LSTM -3-layer LSTM (tied) + continuous cache pointer [28] 53.9 52.8 LSTM+ Dual Channel Class Hierarchy [52] -118.3 LSTM(Large) + cell [53] 76.15 73.87 AWD-FWM [54] 56…”
Section: Transformersmentioning
confidence: 99%