2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA) 2015
DOI: 10.1109/icmla.2015.78
|View full text |Cite
|
Sign up to set email alerts
|

Simplicity of Kmeans Versus Deepness of Deep Learning: A Case of Unsupervised Feature Learning with Limited Data

Abstract: Abstract-We study a bio-detection application as a case study to demonstrate that Kmeans-based unsupervised feature learning can be a simple yet effective alternative to deep learning techniques for small data sets with limited intra-as well as inter-class diversity. We investigate the effect on the classifier performance of data augmentation as well as feature extraction with multiple patch sizes and at different image scales. Our data set includes 1833 images from four different classes of bacteria, each bac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2017
2017
2020
2020

Publication Types

Select...
4
3
3

Relationship

1
9

Authors

Journals

citations
Cited by 23 publications
(12 citation statements)
references
References 7 publications
(6 reference statements)
0
12
0
Order By: Relevance
“…Limited data is a challenge both in getting well-labelled data for supervised learning problems [8] and also continues to be a bigger challenge for low-resource languages [5]. Here we present what the effect of the different augmentation schemes are when labelled data is reduced.…”
Section: Effect Of Augmentation On Less Datamentioning
confidence: 94%
“…Limited data is a challenge both in getting well-labelled data for supervised learning problems [8] and also continues to be a bigger challenge for low-resource languages [5]. Here we present what the effect of the different augmentation schemes are when labelled data is reduced.…”
Section: Effect Of Augmentation On Less Datamentioning
confidence: 94%
“…Deep learning requires huge training datasets to capture features and build the relationships between the set of independent variables and the dependent variables, which can sometimes be difficult to find and tedious to create. This sometimes causes over-fitting triggered by the presence of small datasets, which may drastically influence the recognition capacity of the machine-learning models ( 45 , 46 ). Another disadvantage of deep learning is that it sometimes induces detection latency as a result of the high processing demands ( 47 ).…”
Section: Literature Reviewmentioning
confidence: 99%
“…With the development of unsupervised feature learning techniques [3], deep learning methods proved successful in natural language processing tasks through neural language models [10,33,36]. These models have been used to capture the semantic and syntactic structures of human language [8], and even logical analogies [20], by embedding words as vectors.…”
Section: Related Workmentioning
confidence: 99%