Proceedings of the 2018 SIAM International Conference on Data Mining 2018
DOI: 10.1137/1.9781611975321.58
|View full text |Cite
|
Sign up to set email alerts
|

A Practitioners' Guide to Transfer Learning for Text Classification using Convolutional Neural Networks

Abstract: Transfer Learning (TL) plays a crucial role when a given dataset has insufficient labeled examples to train an accurate model. In such scenarios, the knowledge accumulated within a model pre-trained on a source dataset can be transferred to a target dataset, resulting in the improvement of the target model. Though TL is found to be successful in the realm of imagebased applications, its impact and practical use in Natural Language Processing (NLP) applications is still a subject of research. Due to their hiera… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
20
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 38 publications
(26 citation statements)
references
References 10 publications
0
20
0
Order By: Relevance
“…In such applications, computer vision models pre-trained on a very large but general image data (e.g., ImageNet) are exploited to transfer knowledge to a specialized clinical imaging dataset which is relatively small but sufficient for domain-driven fine-tuning of the general trained model. The success of applying transfer learning on image applications, opened up the possibility to exploit transfer learning in non-clinical NLP applications, such as sentiment classification [42]. However, applying transfer learning of DL models to clinical NLP tasks is still an understudied research topic.…”
Section: Transfer Learningmentioning
confidence: 99%
“…In such applications, computer vision models pre-trained on a very large but general image data (e.g., ImageNet) are exploited to transfer knowledge to a specialized clinical imaging dataset which is relatively small but sufficient for domain-driven fine-tuning of the general trained model. The success of applying transfer learning on image applications, opened up the possibility to exploit transfer learning in non-clinical NLP applications, such as sentiment classification [42]. However, applying transfer learning of DL models to clinical NLP tasks is still an understudied research topic.…”
Section: Transfer Learningmentioning
confidence: 99%
“…This information is used in transfer learning algorithms [45][46][47] to develop knowledge of surrounding rocks and their chemical makeup in upcoming Mars missions (Mars Perseverance Rover, its probes and helicopter) and in future outer and deep space missions (satellites, helicopters and rovers). Further, transfer learning will be used to effectively process source logic networks to target domain and then revises the mapped structure to further improve its accuracy [48].…”
Section: Transfer Learningmentioning
confidence: 99%
“…One way to understand the effect of each of these layers is to finetune or freeze these layers during model transfer and report the best performing model. However, as suggested by [29], the best performance is realized when all layers of a pre-trained model on D S are transferred and the model is led to fine-tune itself using D G . Therefore, we follow the same practice and let the transferred model fine-tune all trainable variables in our model.…”
Section: Transferring Network Layersmentioning
confidence: 99%