Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining 2023
DOI: 10.1145/3580305.3599833
|View full text |Cite
|
Sign up to set email alerts
|

Graph-Aware Language Model Pre-Training on a Large Graph Corpus Can Help Multiple Graph Applications

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
4
1
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(5 citation statements)
references
References 16 publications
0
5
0
Order By: Relevance
“…In this work we propose a model powered by a static layer combining a GNN and an LLM, showing how a synergy between the two architectures can help to combine content and consumption patterns. In a similar fashion, Xie et al [19] in their work show how a graph-aware Language Model framework can help to improve performances on different downstream tasks on large scale industry data. The static layer proposed in our work can be seen as analogous to the pre-training architecture proposed by Xie et al It is worth noting that there are multiple previous efforts combining graphs with LLMs, but the focus has been less on creating a foundation model.…”
Section: Related Workmentioning
confidence: 96%
See 1 more Smart Citation
“…In this work we propose a model powered by a static layer combining a GNN and an LLM, showing how a synergy between the two architectures can help to combine content and consumption patterns. In a similar fashion, Xie et al [19] in their work show how a graph-aware Language Model framework can help to improve performances on different downstream tasks on large scale industry data. The static layer proposed in our work can be seen as analogous to the pre-training architecture proposed by Xie et al It is worth noting that there are multiple previous efforts combining graphs with LLMs, but the focus has been less on creating a foundation model.…”
Section: Related Workmentioning
confidence: 96%
“…On the other hand, graph-based learning models, specifically Graph Neural Networks (GNNs), have emerged as a powerful technology for recommendation systems at scale, becoming a core functionality on different online and social platforms [4,9,19,22]. Moreover, only lately, GNNs have been showing relevant gains also for enabling discovery without loss in accuracy [3].…”
mentioning
confidence: 99%
“…• Pre-training data: Models pre-trained on large amounts of text data to learn language representations can be useful. 78,80 The pre-training data may include various sources of biomedical literature, clinical notes, EHRs, drug labels, and other healthcarerelated text. These datasets can range from millions to billions of tokens.…”
Section: Parameters Used In the Development Of Gpt For Medicinementioning
confidence: 99%
“…Graph-Aware LLM Finetuning SPECTER [51], SciNCL [52], Touchup-G [54], TwHIN-BERT [56], MICoL [59], E2EG [60] LLM as Encoder Optimization One-step TextGNN [77], AdsGNN [78], GNN-LM [66] Two-step GIANT [58], LM-GNN [68], SimTeG [35], GaLM [80] Data Augmentation LLM-GNN [64], TAPE [70], ENG [71] Knowledge Distillation AdsGNN [78], GraD [69] LLM as Aligner Prediction Alignment LTRN [57], GLEM [62] Latent Space Alignment ConGrat [53], GRENADE [55], G2P2 [63], THLM [33] Text-Paired Graphs…”
Section: Graph As Sequencementioning
confidence: 99%
“…They further find that using the efficient fine-tuning method, e.g., LoRA [40] to tune the LLM can alleviate overfitting issues. GaLM [80] explores ways to pretrain the LLM-GNN cascaded architecture. The twostep strategy can effectively alleviate the insufficient training of the LLM which contributes to higher text representation quality but is more computationally expensive and timeconsuming than the one-step training strategy.…”
Section: Two-step Training Means First Adaptingmentioning
confidence: 99%