2019
DOI: 10.3390/e21100988
|View full text |Cite
|
Sign up to set email alerts
|

Combination of Active Learning and Semi-Supervised Learning under a Self-Training Scheme

Abstract: One of the major aspects affecting the performance of the classification algorithms is the amount of labeled data which is available during the training phase. It is widely accepted that the labeling procedure of vast amounts of data is both expensive and time-consuming since it requires the employment of human expertise. For a wide variety of scientific fields, unlabeled examples are easy to collect but hard to handle in a useful manner, thus improving the contained information for a subject dataset. In this … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 17 publications
(12 citation statements)
references
References 62 publications
0
11
0
Order By: Relevance
“…In order to validate our proposed framework, we also present comparisons with a recent state-of-the-art active semi-supervised framework [ 35 ], considering the RF classifier for other biological datasets, such as Haberman, Heart Statlog and Lymphograph [ 54 ] (see Table 9 ). Our framework achieves higher accuracies and requires fewer (much less than 10% of the dataset of) labeled training samples compared to the state of the art one, considering all datasets and active learning strategies.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…In order to validate our proposed framework, we also present comparisons with a recent state-of-the-art active semi-supervised framework [ 35 ], considering the RF classifier for other biological datasets, such as Haberman, Heart Statlog and Lymphograph [ 54 ] (see Table 9 ). Our framework achieves higher accuracies and requires fewer (much less than 10% of the dataset of) labeled training samples compared to the state of the art one, considering all datasets and active learning strategies.…”
Section: Resultsmentioning
confidence: 99%
“…Besides that, in their experiments, the authors presented comparisons between only logistic approaches. Another work [ 35 ] employs a self-training method in which the entropy of unlabeled samples is used in the active learning process, while the semi-supervised learning uses the probability distribution of all possible labels for the samples. Other combinations of active learning and semi-supervised techniques have been proposed in the literature and somewhat successful when applied to distinct contexts, such as: face recognition [ 36 ], diagnosis of intestinal parasites in humans [ 30 ], extraction of protein interaction sentences [ 37 ], unknown and label-scarce classes [ 38 ], sound classification [ 39 ], intrusion detection system [ 40 , 41 ] and textual classification [ 42 ].…”
Section: Introductionmentioning
confidence: 99%
“…It is well known that the configurations of the training set and testing set play important roles in the assessment of the LCZ classification [28,45,59]. Since the ground truth data was usually limited, 10 labeled samples were randomly selected in each class for simulating the insufficiency of labeled samples and the remaining samples were used as a testing set, which is widely used in many semi-supervised researches [60,61]. Also, the final experimental results were the average performance of 10 run outcomes for providing better representativeness.…”
Section: Experimental Descriptionmentioning
confidence: 99%
“…Semisupervised learning is a technique that utilizes unlabeled data to improve the label efficiency. Combining AL with semisupervised learning can increase the label efficiency further [ 16 ]. Graph neural network (GNN) models have achieved state-of-the-art performance in node classification [ 17 ].…”
Section: Introductionmentioning
confidence: 99%