Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2022
DOI: 10.18653/v1/2022.acl-long.52
|View full text |Cite
|
Sign up to set email alerts
|

Early Stopping Based on Unlabeled Samples in Text Classification

Abstract: Early stopping, which is widely used to prevent overfitting, is generally based on a separate validation set. However, in low resource settings, validation-based stopping can be risky because a small validation set may not be sufficiently representative, and the reduction in the number of samples by validation split may result in insufficient samples for training. In this study, we propose an early stopping method that uses unlabeled samples. The proposed method is based on confidence and class distribution si… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 8 publications
0
3
0
Order By: Relevance
“…The number of training iterations was 100, and the model performance was checked using the validation set. We used the early-stopping ( 39 ) method during the training. If the effect has not improved for 10 consecutive rounds, then training is terminated.…”
Section: Resultsmentioning
confidence: 99%
“…The number of training iterations was 100, and the model performance was checked using the validation set. We used the early-stopping ( 39 ) method during the training. If the effect has not improved for 10 consecutive rounds, then training is terminated.…”
Section: Resultsmentioning
confidence: 99%
“…This advantage is useful in low-resource settings. Choi et al [41] proposed an early stopping method based on unlabeled samples (BUS-stop), which performed well, particularly in low and imbalanced data settings. Garg et al [42] also leveraged unlabeled data for early stopping but focused on providing a theoretical perspective on generalization.…”
Section: Related Workmentioning
confidence: 99%
“…The confidences are sorted in the order of size before calculating the similarity. The detailed process is the same as the conf-sim method in BUS-stop [41], except that the confidences are based only on training data.…”
Section: B Non-validation Early Stopping Criteriamentioning
confidence: 99%