Proceedings of the 24th International Conference on Machine Learning 2007
DOI: 10.1145/1273496.1273592
|View full text |Cite
|
Sign up to set email alerts
|

Self-taught learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
70
0
1

Year Published

2012
2012
2023
2023

Publication Types

Select...
5
3
1
1

Relationship

0
10

Authors

Journals

citations
Cited by 1,122 publications
(100 citation statements)
references
References 21 publications
1
70
0
1
Order By: Relevance
“…More specifically, a dictionary could be trained using both sets of labelled and unlabelled data. Related to this is the use of self-taught learning [12] also known as transfer learning from unlabelled data. One could train a large dictionary on a corpus that have come from a different distribution than the target dataset intended for classification.…”
Section: Discussionmentioning
confidence: 99%
“…More specifically, a dictionary could be trained using both sets of labelled and unlabelled data. Related to this is the use of self-taught learning [12] also known as transfer learning from unlabelled data. One could train a large dictionary on a corpus that have come from a different distribution than the target dataset intended for classification.…”
Section: Discussionmentioning
confidence: 99%
“…For target domains with limited example capacities, feature representation seems critical. Ideally, the target domain dataset can well inherit the representational structure transferred from the model trained on the source domain dataset [48,49].…”
Section:  Cross-domain Transfermentioning
confidence: 99%
“…This approach is proposed to be complemented as follows. Weights of the j -th hidden layer after pretraining with auto-encoder can be refined by training network with one hidden layer on the original or a similar problem, using transfer learning concept [7][8][9]. That means, we have to use the weights W , received on the second paragraph of the j -th step of the algorithm for initialization of the network with one hidden layer, an and an output layer size of m , and train the network on , 1 { , }, =1…”
Section: Algorithm Descriptionmentioning
confidence: 99%