2023
DOI: 10.1016/j.jmb.2023.167963
|View full text |Cite
|
Sign up to set email alerts
|

ISPRED-SEQ: Deep Neural Networks and Embeddings for Predicting Interaction Sites in Protein Sequences

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
13
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

2
4

Authors

Journals

citations
Cited by 10 publications
(15 citation statements)
references
References 39 publications
2
13
0
Order By: Relevance
“…Residue-level representations are then concatenated together, leading to vectors of 2304 dimensions for each residue in the sequences. The concatenation of embeddings obtained with different pLMs has been shown to improve the performance in previous works (Manfredi et al, 2022(Manfredi et al, , 2023.…”
Section: Protein Encodingmentioning
confidence: 95%
“…Residue-level representations are then concatenated together, leading to vectors of 2304 dimensions for each residue in the sequences. The concatenation of embeddings obtained with different pLMs has been shown to improve the performance in previous works (Manfredi et al, 2022(Manfredi et al, , 2023.…”
Section: Protein Encodingmentioning
confidence: 95%
“…The next strategy for protein sequence representation is the use of embeddings and encoders. Encoders such as one hot encodings can be used to represent protein sequences, while embeddings such as Word2Vec, FastText, and BERT [7,34,43,44,48,67,88] are commonly used to transform the protein sequence into an interpretable feature for a neural network. The main advantage to this strategy is that it keeps some resemblance of the entire protein sequence while also providing an interpretable feature representation for a neural network [43].…”
Section: Embeddingsmentioning
confidence: 99%
“…However, models that predict local interaction tend to use a combination of statistical, physiochemical representations as well as some representation of the overall protein sequence that captures local features of the protein (e.g., overall fold or domains). As described in the previous sections, protein sequence representations encompass encoding methods such as metric representations, text embeddings, and neural network feature embeddings, but some groups have also leveraged raw protein sequences [39,42,44,67,86,87]. Using unprocessed protein sequences for PPI prediction creates an issue for neural network architectures since most models depend on an input of fixed length.…”
Section: Raw Protein Sequencesmentioning
confidence: 99%
See 2 more Smart Citations