Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1655
|View full text |Cite
|
Sign up to set email alerts
|

Learning with Noisy Labels for Sentence-level Sentiment Classification

Abstract: Deep neural networks (DNNs) can fit (or even over-fit) the training data very well. If a DNN model is trained using data with noisy labels and tested on data with clean labels, the model may perform poorly. This paper studies the problem of learning with noisy labels for sentence-level sentiment classification. We propose a novel DNN model called NETAB (as shorthand for convolutional neural NETworks with AB-networks) to handle noisy labels during training. NETAB consists of two convolutional neural networks, o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
28
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 47 publications
(28 citation statements)
references
References 31 publications
0
28
0
Order By: Relevance
“…To avoid this, it can be combined with label noise handling techniques. This pipeline has been shown to be effective for several NLP tasks (Lange et al, 2019;Paul et al, 2019;Wang et al, 2019;Chen et al, 2019;Mayhew et al, 2019), however, mostly for RNN based approaches. As we have seen in Section 4 that these have a lower baseline performance, we are interested in whether distant supervision is still useful for the better performing transformer models.…”
Section: Distant Supervisionmentioning
confidence: 99%
See 1 more Smart Citation
“…To avoid this, it can be combined with label noise handling techniques. This pipeline has been shown to be effective for several NLP tasks (Lange et al, 2019;Paul et al, 2019;Wang et al, 2019;Chen et al, 2019;Mayhew et al, 2019), however, mostly for RNN based approaches. As we have seen in Section 4 that these have a lower baseline performance, we are interested in whether distant supervision is still useful for the better performing transformer models.…”
Section: Distant Supervisionmentioning
confidence: 99%
“…We use a confusion matrix which is a common approach for handling noisy labels (see, e.g. (Fang and Cohn, 2016;Luo et al, 2017;Lange et al, 2019;Wang et al, 2019)). The confusion matrix models the relationship between the true, clean label of an instance and its corresponding noisy label.…”
Section: E4 Transfer Learningmentioning
confidence: 99%
“…For aspect classification, Karamanolakis et al (2019) create a simple bag-of-words classifier on a list of seed words and train a deep neural network on its weak supervision. Wang et al (2019) use context by transferring a documentlevel sentiment label to all its sentence-level in-stances. Mekala et al (2020) leverage meta-data for text classification and Huber and Carenini (2020) build a discourse-structure dataset using guidance from sentiment annotations.…”
Section: Distant and Weak Supervisionmentioning
confidence: 99%
“…The noise in the labels can also be modeled. A common model is a confusion matrix estimating the relationship between clean and noisy labels (Fang and Cohn, 2016;Luo et al, 2017;Hedderich and Klakow, 2018;Paul et al, 2019;Lange et al, 2019a,c;Wang et al, 2019;Hedderich et al, 2021b). The classifier is no longer trained directly on the noisily-labeled data.…”
Section: Learning With Noisy Labelsmentioning
confidence: 99%
“…The most prominent idea in this line is to estimate the noise transition matrix among labels Goldberger and Ben-Reuven, 2016;Wang et al, 2019;Northcutt et al, 2019) and then use the transition matrices to re-label the instances or adapt the loss functions. Specifically, Wang et al (2019) and Northcutt et al (2019) generate label noise by flipping clean labels based on such noise transition matrices. They are thus not applicable to our weak supervision setting where no clean labels are given.…”
Section: Related Workmentioning
confidence: 99%