2019
DOI: 10.1016/j.jneumeth.2019.108344
|View full text |Cite
|
Sign up to set email alerts
|

Multi-view learning-based data proliferator for boosting classification using highly imbalanced classes

Abstract: Background. Multi-view data representation learning explores the relationship between the views and provides rich complementary information that can improve computer-aided diagnosis. Specifically, existing machine learning methods devised to automate neurological disorder diagnosis using brain data provided new insights into how a particular disorder such as autism spectrum disorder (ASD) alters the brain construct. However, the performance of machine learning methods highly depends on the size of the training… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
21
0
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
7
1
1

Relationship

1
8

Authors

Journals

citations
Cited by 23 publications
(22 citation statements)
references
References 50 publications
0
21
0
1
Order By: Relevance
“…The present study complements a wide array of insightful investigations that have explored the pitfalls and potential solutions for supervised learning with imbalanced data [11,13,46,47,48,17,49,50,32,51,20,52,14,53]. By contrast to some of the previous studies, the present work focuses on insights that are directly relevant to researchers in neuroimaging using standard tools, such as those available through the scikit-learn library.…”
Section: Discussionmentioning
confidence: 96%
See 1 more Smart Citation
“…The present study complements a wide array of insightful investigations that have explored the pitfalls and potential solutions for supervised learning with imbalanced data [11,13,46,47,48,17,49,50,32,51,20,52,14,53]. By contrast to some of the previous studies, the present work focuses on insights that are directly relevant to researchers in neuroimaging using standard tools, such as those available through the scikit-learn library.…”
Section: Discussionmentioning
confidence: 96%
“…However, this comes at the cost of reducing the sample size, increasing the signal-to-noise ratio, which can be detrimental to the classification. Alternatively, one can oversample the minority class by duplicating or interpolating observations [17, 18, 19] (Fig. 1a), though this comes with a higher risk of overfitting and introducing noise [20].…”
Section: Introductionmentioning
confidence: 99%
“…e diversified ensemble learning framework, which finds the best classification algorithm for each individual subdataset, is proposed in the literature [17,18]. Graa and Rekik [19] proposed a multiview learning-based data proliferator (MV-LEAP) that enables the classification of imbalanced multiview representations. Shi et al [20] proposed a general multiple distribution selection method for imbalanced data classification, by proving that traditional classification methods that use single softmax distribution are limited for modeling complex and imbalanced data.…”
Section: Modeling Of High-dimensional and Unbalanced Datamentioning
confidence: 99%
“…However, this overlooks the complementarity and richness of multimodal connectomes and becomes more time-consuming as the number of modalities increases. Another solution is to combine different multimodal connectomes in the given population by simple concatenation and then train a single classifier using the concatenated feature vectors (Dai and He, 2014;Schouten et al, 2016;Raeper et al, 2018;Graa and Rekik, 2019;Corps and Rekik, 2019). However, such approach requires paired samples across modalities, i.e., all MRI modalities should be available for each subject in the population.…”
Section: Introductionmentioning
confidence: 99%