Findings of the Association for Computational Linguistics: EMNLP 2020 2020
DOI: 10.18653/v1/2020.findings-emnlp.101
|View full text |Cite
|
Sign up to set email alerts
|

Using the Past Knowledge to Improve Sentiment Classification

Abstract: This paper studies sentiment classification in the lifelong learning setting that incrementally learns a sequence of sentiment classification tasks. It proposes a new lifelong learning model (called L2PG) that can retain and selectively transfer the knowledge learned in the past to help learn the new task. A key innovation of this proposed model is a novel parameter-gate (p-gate) mechanism that regulates the flow or transfer of the previously learned knowledge to the new task. Specifically, it can selectively … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3

Relationship

4
4

Authors

Journals

citations
Cited by 11 publications
(11 citation statements)
references
References 19 publications
0
11
0
Order By: Relevance
“…Wang et al (2018) used LL for ASC, but improved only the new task and did not deal with CF. Existing CL systems SRK (Lv et al, 2019), KAN (Ke et al, 2020b) and L2PG (Qin et al, 2020) are for document sentiment classification, but not ASC. Ke et al (2020a) also performed transfer in the image domain.…”
Section: Continual Learning Existing Work Has Mainly Focused On Dealing With Catastrophic Forgetting (Cf)mentioning
confidence: 99%
“…Wang et al (2018) used LL for ASC, but improved only the new task and did not deal with CF. Existing CL systems SRK (Lv et al, 2019), KAN (Ke et al, 2020b) and L2PG (Qin et al, 2020) are for document sentiment classification, but not ASC. Ke et al (2020a) also performed transfer in the image domain.…”
Section: Continual Learning Existing Work Has Mainly Focused On Dealing With Catastrophic Forgetting (Cf)mentioning
confidence: 99%
“…L2PG (Qin et al, 2020) uses a neural network but improves only the new task learning for DSC. Wang et al (2018) worked on ASC, but since they improve only the new task learning, they did not deal with CF.…”
Section: Related Workmentioning
confidence: 99%
“…After learning a task, its training data is often discarded (Chen and Liu, 2018). The CL setting is useful when the data privacy is a concern, i.e., the data owners do not want their data used by others (Ke et al, 2020b;Qin et al, 2020;Ke et al, 2021). In such cases, if we want to leverage the knowledge learned in the past to improve the new task learning, CL is appropriate as it shares only the learned model, but not the data.…”
Section: Introductionmentioning
confidence: 99%
“…Some NLP applications have also dealt with CF. For example, CL models have been proposed for sentiment analysis [23,24,37,43], dialogue slot filling [53], language modeling [58,7], language learning [31], sentence embedding [33], machine translation [25], cross-lingual modeling [35], and question answering [12]. A dialogue CL dataset is also reported in [38].…”
Section: Related Workmentioning
confidence: 99%
“…[55] used LL for aspect extraction. [43] and [62] used neural networks for DSC and ASC, respectively. Several papers also studied lifelong topic modeling [5,13].…”
Section: Related Workmentioning
confidence: 99%