Proceedings of the 2018 Conference of the North American Chapter Of the Association for Computational Linguistics: Hu 2018
DOI: 10.18653/v1/n18-1115
|View full text |Cite
|
Sign up to set email alerts
|

The Context-Dependent Additive Recurrent Neural Net

Abstract: Contextual sequence mapping is one of the fundamental problems in Natural Language Processing. Instead of relying solely on the information presented in a text, the learning agents have access to a strong external signal given to assist the learning process. In this paper, we propose a novel family of Recurrent Neural Network unit: the Context-dependent Additive Recurrent Neural Network (CARNN) that is designed specifically to leverage this external signal. The experimental results on public datasets in the di… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 19 publications
(7 citation statements)
references
References 16 publications
0
7
0
Order By: Relevance
“…with increasingly challenging types of questions. For the task of AS2, initial efforts embedded the question and candidates using CNNs (Severyn and Moschitti, 2015), weight aligned networks (Shen et al, 2017;Tran et al, 2018;Tay et al, 2018) and compare-aggregate architectures (Wang and Jiang, 2016;Bian et al, 2017;Yoon et al, 2019). Recent progress has stemmed from the application of transformer models for performing AS2 (Garg et al, 2020;Han et al, 2021;Lauriola and Moschitti, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…with increasingly challenging types of questions. For the task of AS2, initial efforts embedded the question and candidates using CNNs (Severyn and Moschitti, 2015), weight aligned networks (Shen et al, 2017;Tran et al, 2018;Tay et al, 2018) and compare-aggregate architectures (Wang and Jiang, 2016;Bian et al, 2017;Yoon et al, 2019). Recent progress has stemmed from the application of transformer models for performing AS2 (Garg et al, 2020;Han et al, 2021;Lauriola and Moschitti, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…Answer Sentence Selection (AS2) In the last few years, several approaches have been proposed for AS2. For example, Severyn and Moschitti (2015) applied CNN to create question and answer representations, while others proposed interweighted alignment networks (Shen et al, 2017;Tran et al, 2018;Tay et al, 2018). The use of compare and aggregate architectures has also been extensively evaluated (Wang and Jiang, 2016;Bian et al, 2017;Yoon et al, 2019).…”
Section: Related Workmentioning
confidence: 99%
“…The forward LSTM processes the question from left to right, it outputs sequences − → h t ; while the backward LSTM processes the question in the reverse direction, it outputs sequences ← − h t . Both the forward and backward layer outputs are computed by using the standard LSTM updating equations, Equations (1) - (6). The output at each 2 http://www.ltp-cloud.com/demo/ time step is the concatenation of the two output vectors from both directions, which is calculated by using the following equation:…”
Section: Attention-based Bi-lstmmentioning
confidence: 99%
“…In factoid question answering, the word embedding is trained based on word2vec [45] and Chinese Wikipedia Dump 6 . The dimension of word vectors is set to 300.…”
Section: Evaluation Metrics and Experimental Settingsmentioning
confidence: 99%
See 1 more Smart Citation