Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.50
|View full text |Cite
|
Sign up to set email alerts
|

Bayes-enhanced Lifelong Attention Networks for Sentiment Classification

Abstract: The classic deep learning paradigm learns a model from the training data of a single task and the learned model is also tested on the same task. This paper studies the problem of learning a sequence of tasks (sentiment classification tasks in our case). After each sentiment classification task is learned, its knowledge is retained to help future task learning. Following this setting, we explore attention neural networks and propose a Bayes-enhanced Lifelong Attention Network (BLAN). The key idea is to exploit … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 40 publications
0
2
0
Order By: Relevance
“…Wang et al [191] impose the idea of lifelong learning into the ASC task and propose a novel lifelong learning approach based on memory networks. Recent lifelong sentiment analysis studies begin to investigate the catastrophic forgetting issue during the sequential learning [192,193,194,195], instead of simply studying it as an extension of crossdomain sentiment analysis for knowledge accumulation. However, existing studies mainly focus on domain incremental learning for the ASC task [194,195], where all tasks sharing the same fixed label classes (e.g., positive, negative, and neutral) and no task information is required.…”
Section: Lifelong Absamentioning
confidence: 99%
“…Wang et al [191] impose the idea of lifelong learning into the ASC task and propose a novel lifelong learning approach based on memory networks. Recent lifelong sentiment analysis studies begin to investigate the catastrophic forgetting issue during the sequential learning [192,193,194,195], instead of simply studying it as an extension of crossdomain sentiment analysis for knowledge accumulation. However, existing studies mainly focus on domain incremental learning for the ASC task [194,195], where all tasks sharing the same fixed label classes (e.g., positive, negative, and neutral) and no task information is required.…”
Section: Lifelong Absamentioning
confidence: 99%
“…Additionally, they used a word grid graph as input to retain multigranularity information, serving as a complement to pre-trained language models. Wang et al introduced a novel model called Dabert, which improves sentence semantic modeling through distance-aware self-attention and multi-level matching [6]. The model considers the importance of word distances and captures interactions between sentences.…”
Section: Introductionmentioning
confidence: 99%