2021
DOI: 10.1101/2021.09.03.458869
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Protein embeddings and deep learning predict binding residues for various ligand classes

Abstract: One important aspect of protein function is the binding of proteins to ligands, including small molecules, metal ions, and macromolecules such as DNA or RNA. Despite decades of experimental progress many binding sites remain obscure. Here, we proposed bindEmbed21, a method predicting whether a protein residue binds to metal ions, nucleic acids, or small molecules. The Artificial Intelligence (AI)-based method exclusively uses embeddings from the Transformer-based protein Language Model ProtT5 as input. Using o… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

6
2

Authors

Journals

citations
Cited by 11 publications
(12 citation statements)
references
References 58 publications
(166 reference statements)
0
12
0
Order By: Relevance
“…5 , Fig. 6 ), although for SAV effect predictions, embedding-based methods are still not yet outperforming the MSA-based SOTA as for other prediction tasks (Elnaggar et al 2021 ; Littmann et al 2021a , b , c ; Stärk et al 2021 ). Embedding-based predictions are blazingly fast, thereby they save computing, and ultimately energy resources when applied to daily sequence analysis.…”
Section: Discussionmentioning
confidence: 95%
See 1 more Smart Citation
“…5 , Fig. 6 ), although for SAV effect predictions, embedding-based methods are still not yet outperforming the MSA-based SOTA as for other prediction tasks (Elnaggar et al 2021 ; Littmann et al 2021a , b , c ; Stärk et al 2021 ). Embedding-based predictions are blazingly fast, thereby they save computing, and ultimately energy resources when applied to daily sequence analysis.…”
Section: Discussionmentioning
confidence: 95%
“…1 in (Elnaggar et al 2021 )]. Embeddings have succeeded as exclusive input to predicting secondary structure and subcellular location at performance levels almost reaching (Alley et al 2019 ; Heinzinger et al 2019 ; Rives et al 2021 ) or even exceeding (Elnaggar et al 2021 ; Littmann et al 2021c ; Stärk et al 2021 ) state-of-the-art (SOTA) methods using EI from MSAs as input. Embeddings even succeed in substituting sequence similarity for homology-based annotation transfer (Littmann et al 2021a , b ) and in predicting the effect of mutations on protein–protein interactions (Zhou et al 2020 ).…”
Section: Introductionmentioning
confidence: 99%
“…In their simplest form, embeddings mirror the last "hidden" states/values of pLMs. In analogy to NLPs implicitly learning grammar, embeddings from pLMs capture some aspects of the language of life as written in protein sequences (Alley et al, 2019;Heinzinger et al, 2019;Ofer et al, 2021;Rives et al, 2021), which suffices as exclusive input to many methods predicting aspects of protein structure and function (Asgari and Mofrad, 2015;Alley et al, 2019;Heinzinger et al, 2019;Littmann et al, 2021a;Littmann et al, 2021b;Littmann et al, 2021c;Elnaggar et al, 2021;Heinzinger et al, 2021;Marquet et al, 2021;Rives et al, 2021).…”
Section: Many Prediction Methods Availablementioning
confidence: 99%
“…Although falling substantially short of AlphaFold2 (Jumper et al, 2021). Per-protein embeddings outperform the best MSA-based methods in the prediction of sub-cellular location (Staerk et al, 2021), signal peptides (Teufel et al, 2021) and binding residues (Littmann et al, 2021c).…”
Section: Protein Language Models Capture Crucial Constraintsmentioning
confidence: 95%