2018
DOI: 10.48550/arxiv.1806.01547
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Semi-Supervised Clustering with Neural Networks

Abstract: Clustering using neural networks has recently demonstrated promising performance in machine learning and computer vision applications. However, the performance of current approaches is limited either by unsupervised learning or their dependence on large set of labeled data samples. In this paper, we propose ClusterNet that uses pairwise semantic constraints from very few labeled data samples (< 5% of total data) and exploits the abundant unlabeled data to drive the clustering approach. We define a new loss fun… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
7
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 15 publications
0
7
0
Order By: Relevance
“…The authors of [27] used a KL-divergence-based loss to train a DNN to predict cluster distribution from pairwise relations; one limitation of that method is its inability to use unlabeled data. Other works [28,29,30] used autoencoders with reconstruction losses to exploit inner characteristics of unlabeled data.…”
Section: Related Workmentioning
confidence: 99%
“…The authors of [27] used a KL-divergence-based loss to train a DNN to predict cluster distribution from pairwise relations; one limitation of that method is its inability to use unlabeled data. Other works [28,29,30] used autoencoders with reconstruction losses to exploit inner characteristics of unlabeled data.…”
Section: Related Workmentioning
confidence: 99%
“…Representation Learning & Clustering Our approach relies on estimating unknown subclass labels by clustering a feature representation of the data. Techniques for learning semantically useful image features include autoencoder-based methods [32,46], the use of unsupervised auxiliary tasks [2,9], and pretraining on massive datasets [27]. Such features may be used for unsupervised identification of classes, either using clustering techniques [6] or an end-to-end approach [25,18].…”
Section: Related Workmentioning
confidence: 99%
“…This has been an area of substantial recent activity in machine learning, and has provided several important conclusions upon which we build in our work. The work of [32] and [46], for instance, demonstrate the utility of a simple autoencoded representation for performing unsupervised clustering in the feature space of a trained model. While the purpose of these works is often to show that deep clustering can be competitive with semi-supervised learning techniques, the mechanics of clustering in model feature space explored by these works are important for our present study.…”
Section: Neural Representation Clusteringmentioning
confidence: 99%
“…Clustering The problem of clustering can be broadly defined as a label assignment task where data points with similar features are to be assigned the same label. Recently, deep neural networks have been utilized to perform this task in a supervised or semi-supervised [25,33] and unsupervised [4,23,37] setting. Similar to our work, some prior research also explores a voting mechanism [3,9,15,25] for clustering.…”
Section: Related Workmentioning
confidence: 99%