2023
DOI: 10.1016/j.csbj.2022.11.014
|View full text |Cite
|
Sign up to set email alerts
|

From sequence to function through structure: Deep learning for protein design

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
45
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3
2
1

Relationship

1
9

Authors

Journals

citations
Cited by 55 publications
(49 citation statements)
references
References 122 publications
0
45
0
Order By: Relevance
“…More specifically, we used a recently published pipeline that combines a variety of protein property predictors that all rely on the pLM ProtT5 to encode protein sequences. This enables efficient processing as the heavy-lifting of generating embeddings needs to be performed only once while the predictors trained add only insignificant computational overhead allowing to predict thousands of proteins within few minutes [69]. Here, we mainly focused on predicting whether a protein is secreted or not using a method called LightAttention and on predicting transmembrane regions using the method TMbed [70].…”
Section: Methodsmentioning
confidence: 99%
“…More specifically, we used a recently published pipeline that combines a variety of protein property predictors that all rely on the pLM ProtT5 to encode protein sequences. This enables efficient processing as the heavy-lifting of generating embeddings needs to be performed only once while the predictors trained add only insignificant computational overhead allowing to predict thousands of proteins within few minutes [69]. Here, we mainly focused on predicting whether a protein is secreted or not using a method called LightAttention and on predicting transmembrane regions using the method TMbed [70].…”
Section: Methodsmentioning
confidence: 99%
“…This has been followed by a boom in end-to-end learning approaches on proteins sequences for function prediction, as well as on protein structures for generating designed protein sequences. See [223] for a recent review.…”
Section: Challenges and Future Directionsmentioning
confidence: 99%
“…They have also been employed to understand the language of nucleic acids (DNA/RNA) and proteins. Protein language models (PLMs), using amino acid alphabets, are the most heavily investigated biological LLMs (Elnaggar et al, 2022, Rives et al, 2021), with demonstrated success in downstream tasks such as protein function prediction (Unsal et al, 2022) and engineering (Ferruz et al, 2022). Nucleotide LLMs, using DNA/RNA alphabets, are still understudied (Avsec et al, 2021).…”
Section: Current State Of the Artmentioning
confidence: 99%