2018
DOI: 10.1007/978-3-319-73618-1_17
|View full text |Cite
|
Sign up to set email alerts
|

Shortcut Sequence Tagging

Abstract: Deep stacked RNNs are usually hard to train. Adding shortcut connections across different layers is a common way to ease the training of stacked networks. However, extra shortcuts make the recurrent step more complicated. To simply the stacked architecture, we propose a framework called shortcut block, which is a marriage of the gating mechanism and shortcuts, while discarding the selfconnected part in LSTM cell. We present extensive empirical experiments showing that this design makes training easy and improv… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 25 publications
0
2
0
Order By: Relevance
“…Clark and Curran (2007) propose C&C tagger which uses a log-linear model to build the supertagger. Recent works have applied neural networks to supertagging (Xu, Auli, and Clark 2015;Vaswani et al 2016;Wu, Zhang, and Zong 2017b). These works perform a multi-class classification on pre-defined category sets and they can't capture the inside connections between categories because categories are independent of each other.…”
Section: Related Workmentioning
confidence: 99%
“…Clark and Curran (2007) propose C&C tagger which uses a log-linear model to build the supertagger. Recent works have applied neural networks to supertagging (Xu, Auli, and Clark 2015;Vaswani et al 2016;Wu, Zhang, and Zong 2017b). These works perform a multi-class classification on pre-defined category sets and they can't capture the inside connections between categories because categories are independent of each other.…”
Section: Related Workmentioning
confidence: 99%
“…propose C&C tagger which uses a log-linear model to build the supertagger. Recent works have applied neural networks to supertagging (Xu, Auli, and Clark 2015;Vaswani et al 2016;Wu, Zhang, and Zong 2017b). These works perform a multi-class classification on pre-defined category sets and they can't capture the inside connections between categories because categories are independent of each other.…”
Section: Related Workmentioning
confidence: 99%