2023
DOI: 10.1101/2023.10.01.560349
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

SaProt: Protein Language Modeling with Structure-aware Vocabulary

Jin Su,
Chenchen Han,
Yuyang Zhou
et al.

Abstract: Large-scale protein language models (PLMs), such as the ESM family, have achieved remarkable performance in various downstream tasks related to protein structure and function by undergoing unsupervised training on residue sequences. They have become essential tools for researchers and practitioners in biology. However, a limitation of vanilla PLMs is their lack ofexplicitconsideration for protein structure information, which suggests the potential for further improvement. Motivated by this, we introduce the co… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
24
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 43 publications
(39 citation statements)
references
References 52 publications
0
24
0
Order By: Relevance
“…Many pretraining tasks still aim to reconstruct natural sequences (He et al, 2021; Notin et al, 2022; Tan et al, 2023; Ma et al, 2023) and so are also likely to primarily learn coevolutionary patterns. Other tasks use structure as an additional input or target, but they generally make only modest improvements on function prediction tasks (Mansoor et al, 2021; Wang et al, 2022; Yang et al, 2023; Su et al, 2023). Supporting the assertion that learning to predict structure may not improve function prediction, Hu et al (2022) show that transfer learning using the AlphaFold2 (Jumper et al, 2021) structure module is less effective for function prediction than transferring PLMs.…”
Section: Discussionmentioning
confidence: 99%
“…Many pretraining tasks still aim to reconstruct natural sequences (He et al, 2021; Notin et al, 2022; Tan et al, 2023; Ma et al, 2023) and so are also likely to primarily learn coevolutionary patterns. Other tasks use structure as an additional input or target, but they generally make only modest improvements on function prediction tasks (Mansoor et al, 2021; Wang et al, 2022; Yang et al, 2023; Su et al, 2023). Supporting the assertion that learning to predict structure may not improve function prediction, Hu et al (2022) show that transfer learning using the AlphaFold2 (Jumper et al, 2021) structure module is less effective for function prediction than transferring PLMs.…”
Section: Discussionmentioning
confidence: 99%
“…After the transformative progress brought by deep learning to protein structure prediction [2][3][4][5], predicting protein complex structure and ligand binding sites is fast advancing with AFM and related methods, but also with other deep learning models based on structural representations [73][74][75][76]. Combining the latter [77,78] and, more generally, structural information [79] with the power of sequence-based language models is starting to bring even further progress.…”
Section: Discussionmentioning
confidence: 99%
“…Deep learning tools trained on structural information are likely superior predictors in tasks where spatial features are pre-eminent, as in the case of protein-protein interactions (Wang et al 2022, Heinzinger et al 2023, Shu et al 2023, Su et al 2023). AlphaFold2 and ProteinMPNN, a model developed around a message-passing neural network, have shown potential in predicting fitness variations upon residue substitution (Dauparas et al 2022, Brown et al 2023, Reeves and Kalyaanamoorthy 2023).…”
Section: Discussionmentioning
confidence: 99%