2019
DOI: 10.1038/s41417-019-0105-y
|View full text |Cite
|
Sign up to set email alerts
|

Identification of leukemia stem cell expression signatures through Monte Carlo feature selection strategy and support vector machine

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
36
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
9

Relationship

1
8

Authors

Journals

citations
Cited by 45 publications
(37 citation statements)
references
References 97 publications
0
36
0
Order By: Relevance
“…Classification Algorithm. Two robust machine learning techniques, i.e., SVM and RF, are applied to perform the prediction of DBPs, which have been widely used for many classification tasks in the field of computational biology [43][44][45][46]. SVM is an outstanding classification method that is used to deal with a binary pattern recognition problem [47].…”
Section: Featurementioning
confidence: 99%
“…Classification Algorithm. Two robust machine learning techniques, i.e., SVM and RF, are applied to perform the prediction of DBPs, which have been widely used for many classification tasks in the field of computational biology [43][44][45][46]. SVM is an outstanding classification method that is used to deal with a binary pattern recognition problem [47].…”
Section: Featurementioning
confidence: 99%
“…After the irrelevant features were removed, the relevant methylation and expression features were ranked based on their importance evaluated with MCFS (Monte Carlo Feature Selection) (Draminski et al, 2008). The MCFS was a widely used method to rank features based on classification trees (Chen et al, , 2019Pan et al, 2018Pan et al, , 2019aLi et al, 2019). First, for the d features, we selected s subsets and each subset included m features (m was much smaller than d).…”
Section: Evaluate the Importance Of Relevant Methylation And Expressimentioning
confidence: 99%
“…The second stage was to determine the number of selected genes using the IFS method (Chen et al, 2018b;Chen et al, 2019b;Chen et al, 2019c;Chen et al, 2019d;Chen et al, 2019f;Li et al, 2019a;Pan et al, 2019a;Pan et al, 2019b;). To do so, 200 classifiers were constructed using top 1, top 2, top 200 genes.…”
Section: Two Stage Feature Selection Approachmentioning
confidence: 99%
“…We tried several different classifiers: (1) SVM (Support Vector Machine) (Jiang et al, 2019;Yan et al, 2019;Chen et al, 2019a;Li et al, 2019a;Pan et al, 2019a;Wang and Huang, 2019b;Chen et al, 2019d), (2) 1NN (1 Nearest Neighbor) (Lei et al, 2013;Chen et al, 2016;Wang et al, 2017a), (3) 3NN (3 Nearest Neighbors), (4) 5NN (5 Nearest Neighbors), (5) Decision Tree (DT) (Huang et al, 2008;Huang et al, 2011;Chen et al, 2015), (6) Neural Network (NN) (Liu et al, 2017;Pan et al, 2018;Chen et al, 2019e). The function svm from R package e1071, function knn from R package class, function rpart from R package rpart, function nnet from R package nnet were used to apply these classification algorithms.…”
Section: Two Stage Feature Selection Approachmentioning
confidence: 99%