2018
DOI: 10.1016/j.gpb.2018.08.004
|View full text |Cite
|
Sign up to set email alerts
|

Integration of A Deep Learning Classifier with A Random Forest Approach for Predicting Malonylation Sites

Abstract: As a newly-identified protein post-translational modification, malonylation is involved in a variety of biological functions. Recognizing malonylation sites in substrates represents an initial but crucial step in elucidating the molecular mechanisms underlying protein malonylation. In this study, we constructed a deep learning (DL) network classifier based on long short-term memory (LSTM) with word embedding (LSTMWE) for the prediction of mammalian malonylation sites. LSTMWE performs better than traditional cl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
56
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3
2
2

Relationship

2
5

Authors

Journals

citations
Cited by 71 publications
(56 citation statements)
references
References 42 publications
(58 reference statements)
0
56
0
Order By: Relevance
“…The EAAC encoding [26,[32][33][34] introduces a fixed-length sliding window based on the encoding of Amino Acid Composition (AAC), which calculates the frequency of each type of amino acids in a protein or peptide sequence [35]. EAAC is calculated by continuously sliding a fixed-length sequence window (using the default value as 5) from the N-terminus to the C-terminus of each peptide.…”
Section: Enhanced Amino Acid Composition (Eaac)mentioning
confidence: 99%
“…The EAAC encoding [26,[32][33][34] introduces a fixed-length sliding window based on the encoding of Amino Acid Composition (AAC), which calculates the frequency of each type of amino acids in a protein or peptide sequence [35]. EAAC is calculated by continuously sliding a fixed-length sequence window (using the default value as 5) from the N-terminus to the C-terminus of each peptide.…”
Section: Enhanced Amino Acid Composition (Eaac)mentioning
confidence: 99%
“…Khib sites were considered positives whereas the remaining K sites were taken as negatives. We further estimated the potential redundancy of the positive sites by extracting the peptide segment of seven residues with the Khib site in the center and count the number of unique segments [20,25]. The number (9,444) through ten-fold cross-validation ( Fig.…”
Section: Dataset Collectionmentioning
confidence: 99%
“…The DL models included a Gated Recurrent Unit (GRU) model with the word-embedding encoding approach dubbed GRUWE and two CNN models with the one-hot and word-embedding encoding approaches named CNNOH and CNNWE, respectively. Both encoding methods are common in the DL algorithms [20,25].…”
Section: Cnnoh Showed Superior Performancementioning
confidence: 99%
See 2 more Smart Citations