2018
DOI: 10.48550/arxiv.1811.08600
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Contextualized Non-local Neural Networks for Sequence Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
4

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 0 publications
0
4
0
Order By: Relevance
“…(Norcliffe-Brown et al, 2018) dynamically construct a graph which contains all the visual objects appearing in an image. In parallel to our work, (Liu et al, 2018) also dynamically construct a graph of all words from free text.…”
Section: Graph Neural Networkmentioning
confidence: 99%
“…(Norcliffe-Brown et al, 2018) dynamically construct a graph which contains all the visual objects appearing in an image. In parallel to our work, (Liu et al, 2018) also dynamically construct a graph of all words from free text.…”
Section: Graph Neural Networkmentioning
confidence: 99%
“…The success of Transformer has raised a large body of follow-up work. Therefore, some Transformer variations are also proposed, such as GPT (Radford et al, 2018), BERT (Devlin et al, 2018), Transformer-XL (Dai et al, 2019) , Universal Transformer (Dehghani et al, 2018) and CN 3 (Liu et al, 2018a).…”
Section: Modelling Non-local Compositionalitymentioning
confidence: 99%
“…GNN for NLP: Recently, there is considerable amount of interest in applying GNN to NLP tasks and great success has been achieved. For example, in neural machine translation, GNN has been employed to integrate syntactic and semantic information into encoders (Bastings et al, 2017;Marcheggiani et al, 2018); applied GNN to relation extraction over pruned dependency trees; the study by Yao et al (2018) employed GNN over a heterogeneous graph to do text classification, which inspires our idea of the HDE graph; Liu et al (2018) proposed a new contextualized neural network for sequence learning by leveraging various types of non-local contextual information in the form of information passing over GNN. These studies are related to our work in the sense that we both use GNN to improve the information interaction over long context or across documents.…”
Section: Multi-hop Rcmentioning
confidence: 99%