2019
DOI: 10.1609/aaai.v33i01.33016762
|View full text |Cite
|
Sign up to set email alerts
|

Contextualized Non-Local Neural Networks for Sequence Learning

Abstract: Recently, a large number of neural mechanisms and models have been proposed for sequence learning, of which selfattention, as exemplified by the Transformer model, and graph neural networks (GNNs) have attracted much attention. In this paper, we propose an approach that combines and draws on the complementary strengths of these two methods. Specifically, we propose contextualized non-local neural networks (CN 3 ), which can both dynamically construct a task-specific structure of a sentence and leverage rich lo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
15
1

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 37 publications
(16 citation statements)
references
References 2 publications
(2 reference statements)
0
15
1
Order By: Relevance
“…We can conclude that sentence-context information is not significantly helpful for disease mention normalization. This is not consistent with the descovery in previous work, in which sentence information has been shown useful for other NLP tasks [45], [46].…”
Section: B Effect Of Sentence-context Informationcontrasting
confidence: 94%
“…We can conclude that sentence-context information is not significantly helpful for disease mention normalization. This is not consistent with the descovery in previous work, in which sentence information has been shown useful for other NLP tasks [45], [46].…”
Section: B Effect Of Sentence-context Informationcontrasting
confidence: 94%
“…The success of Transformer has raised a large body of follow-up work. Therefore, some Transformer variations are also proposed, such as GPT (Radford et al, 2018), BERT (Devlin et al, 2018), Transformer-XL (Dai et al, 2019) , Universal Transformer (Dehghani et al, 2018) and CN 3 (Liu et al, 2018a).…”
Section: Related Workmentioning
confidence: 99%
“…GNN for NLP: Recently, there is considerable amount of interest in applying GNN to NLP tasks and great success has been achieved. For example, in neural machine translation, GNN has been employed to integrate syntactic and semantic information into encoders (Bastings et al, 2017;Marcheggiani et al, 2018); applied GNN to relation extraction over pruned dependency trees; the study by Yao et al (2018) employed GNN over a heterogeneous graph to do text classification, which inspires our idea of the HDE graph; Liu et al (2018) proposed a new contextualized neural network for sequence learning by leveraging various types of non-local contextual information in the form of information passing over GNN. These studies are related to our work in the sense that we both use GNN to improve the information interaction over long context or across documents.…”
Section: Related Workmentioning
confidence: 99%