2023
DOI: 10.1038/s41467-023-37958-z
|View full text |Cite
|
Sign up to set email alerts
|

Latent generative landscapes as maps of functional diversity in protein sequence space

Abstract: Variational autoencoders are unsupervised learning models with generative capabilities, when applied to protein data, they classify sequences by phylogeny and generate de novo sequences which preserve statistical properties of protein composition. While previous studies focus on clustering and generative features, here, we evaluate the underlying latent manifold in which sequence information is embedded. To investigate properties of the latent manifold, we utilize direct coupling analysis and a Potts Hamiltoni… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 14 publications
(16 citation statements)
references
References 95 publications
0
7
0
Order By: Relevance
“…In total, 21 interpretable features were used in the models, including properties derived from protein sequences, structures, networks, and gene mutational constraints (Badonyi and Marsh 2023). Additionally, 20 language model-based embeddings were also included, which are thought to represent protein function in their latent space [41]. As a measure of feature importance, we calculated the loss in AUROC relative to the full model ( Figure 3 ).…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…In total, 21 interpretable features were used in the models, including properties derived from protein sequences, structures, networks, and gene mutational constraints (Badonyi and Marsh 2023). Additionally, 20 language model-based embeddings were also included, which are thought to represent protein function in their latent space [41]. As a measure of feature importance, we calculated the loss in AUROC relative to the full model ( Figure 3 ).…”
Section: Resultsmentioning
confidence: 99%
“…Additionally, 20 language model-based embeddings were also included, which are thought to represent protein function in their latent space [41]. As a measure of feature importance, we calculated the loss in AUROC relative to the full model (Figure 3).…”
Section: Global and Local Feature Importance Evaluationmentioning
confidence: 99%
“…Formally, H can be used to compute the probability of finding a particular sequence in the input protein family. H serves as a proxy for protein fitness and has been described as a measure of the “typicality” of a protein sequence within its family, with more negative values signifying more family like or typical sequences . Coevolutionary information-based scoring has also been shown to predict specificity between histidine kinases and response regulators, compatibility between DNA recognition and allosteric response modules in LacI-type transcription inhibitors, folding kinetics, and mutational phenotypes in protein–RNA complexes …”
Section: Resultsmentioning
confidence: 99%
“…219 In addition, by exploring the latent manifold underlying the sequence information, we can uncover dependencies that may not be readily apparent in the raw latent space embeddings. 220 Despite its advantages, MSAs also have some drawbacks. First, it can be difficult to create an MSA that contains enough evolutionarily relevant sequences to establish strong patterns at key amino acid positions.…”
Section: Supervised Learning To Predict the Effects Of Mutationsmentioning
confidence: 99%
“…Building on these findings, the geometric structure of a latent space was recently used to guide the design of a haloalkane dehalogenase . In addition, by exploring the latent manifold underlying the sequence information, we can uncover dependencies that may not be readily apparent in the raw latent space embeddings …”
Section: Protein Engineering Tasks Solved By Machine Learningmentioning
confidence: 99%