Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 2021
DOI: 10.18653/v1/2021.findings-acl.126
|View full text |Cite
|
Sign up to set email alerts
|

BertGCN: Transductive Text Classification by Combining GNN and BERT

Abstract: In this work, we propose BertGCN, a model that combines large scale pretraining and transductive learning for text classification. Bert-GCN constructs a heterogeneous graph over the dataset and represents documents as nodes using BERT representations. By jointly training the BERT and GCN modules within Bert-GCN, the proposed model is able to leverage the advantages of both worlds: large-scale pretraining which takes the advantage of the massive amount of raw data and transductive learning which jointly learns … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
40
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
2
1

Relationship

1
9

Authors

Journals

citations
Cited by 100 publications
(46 citation statements)
references
References 46 publications
1
40
0
Order By: Relevance
“…On this account, it is easy to see how graph architectures can also be integrated with deep language models. BertGCN [100], for example, trains a GCN jointly with a BERT-like model, in order to leverage the advantages of both pre-trained language models and graph-based approaches. Document nodes are initialised through BERT-style embeddings and updated iteratively by the GCN layers.…”
Section: Successful Approachesmentioning
confidence: 99%
“…On this account, it is easy to see how graph architectures can also be integrated with deep language models. BertGCN [100], for example, trains a GCN jointly with a BERT-like model, in order to leverage the advantages of both pre-trained language models and graph-based approaches. Document nodes are initialised through BERT-style embeddings and updated iteratively by the GCN layers.…”
Section: Successful Approachesmentioning
confidence: 99%
“…Lee et al [13] adopted a perplexity-based approach in the few-shot learning, which assumes that the given claim may be fake if the corresponding perplexity score from evidence-conditioned language models is high. BertGCN [14] is proposed by integrating the advantages of large-scale pre-trained models and graph neural networks for fake news detection, which is able to learn the representations from the massive amount of pre-trained data and the label influence through the propagation. MCAN [6] adopts a large-scale pre-trained NLP model and a pre-trained computer vision (CV) model for extracting features from text and images, respectively.…”
Section: Fake News Detectionmentioning
confidence: 99%
“…GNN have gained popularity due to their powerful expressive ability, and they are also used to solve the problem of text classification [31][32][33][34].…”
Section: Short Text Classification Based Gcnmentioning
confidence: 99%