Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2021
DOI: 10.18653/v1/2021.naacl-main.378
|View full text |Cite
|
Sign up to set email alerts
|

Adapting BERT for Continual Learning of a Sequence of Aspect Sentiment Classification Tasks

Abstract: This paper studies continual learning (CL) of a sequence of aspect sentiment classification (ASC) tasks. Although some CL techniques have been proposed for document sentiment classification, we are not aware of any CL work on ASC. A CL system that incrementally learns a sequence of ASC tasks should address the following two issues: (1) transfer knowledge learned from previous tasks to the new task to help it learn a better model, and (2) maintain the performance of the models for previous tasks so that they ar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
40
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
2

Relationship

2
6

Authors

Journals

citations
Cited by 53 publications
(60 citation statements)
references
References 31 publications
0
40
0
Order By: Relevance
“…However, after three epochs RULEBERT is also at 0.0, 3 i.e., it started to unlearn what it had learned at pre-fine-tuning (Kirkpatrick et al, 2017;Kemker et al, 2018;Biesialska et al, 2020). Learning a new task often leads to such catastrophic forgetting (Ke et al, 2021). While there are ways to alleviate this (Ke et al, 2021), this is beyond the scope of this paper.…”
Section: Resultsmentioning
confidence: 96%
See 1 more Smart Citation
“…However, after three epochs RULEBERT is also at 0.0, 3 i.e., it started to unlearn what it had learned at pre-fine-tuning (Kirkpatrick et al, 2017;Kemker et al, 2018;Biesialska et al, 2020). Learning a new task often leads to such catastrophic forgetting (Ke et al, 2021). While there are ways to alleviate this (Ke et al, 2021), this is beyond the scope of this paper.…”
Section: Resultsmentioning
confidence: 96%
“…Learning a new task often leads to such catastrophic forgetting (Ke et al, 2021). While there are ways to alleviate this (Ke et al, 2021), this is beyond the scope of this paper.…”
Section: Resultsmentioning
confidence: 98%
“…Existing CL systems SRK (Lv et al, 2019) and KAN (Ke et al, 2020b) are for DSC in the TIL setting, not for ASC. B-CL (Ke et al, 2021) is the first CL system for ASC. It also uses the idea of Adapter-BERT in (Houlsby et al, 2019) and is based on Capsule Network.…”
Section: Related Workmentioning
confidence: 99%
“…After learning a task, its training data is often discarded (Chen and Liu, 2018). The CL setting is useful when the data privacy is a concern, i.e., the data owners do not want their data used by others (Ke et al, 2020b;Qin et al, 2020;Ke et al, 2021). In such cases, if we want to leverage the knowledge learned in the past to improve the new task learning, CL is appropriate as it shares only the learned model, but not the data.…”
Section: Introductionmentioning
confidence: 99%
“…Another observation about the current CL research is that most techniques do not use pre-trained models. But such pre-trained models or feature extractors can significantly improve the CL performance [18,24]. An important question is how to make the best use of pre-trained models in CL.…”
Section: Introductionmentioning
confidence: 99%