Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2021
DOI: 10.18653/v1/2021.naacl-main.308
|View full text |Cite
|
Sign up to set email alerts
|

Pruning-then-Expanding Model for Domain Adaptation of Neural Machine Translation

Abstract: Domain Adaptation is widely used in practical applications of neural machine translation, which aims to achieve good performance on both general domain and in-domain data. However, the existing methods for domain adaptation usually suffer from catastrophic forgetting, large domain divergence, and model explosion. To address these three problems, we propose a method of "divide and conquer" which is based on the importance of neurons or parameters for the translation model. In this method, we first prune the mod… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 17 publications
(16 citation statements)
references
References 34 publications
0
12
0
Order By: Relevance
“…Identifying the salient neurons with respect to a domain can be effectively used for domain adaptation and generalization. Gu et al (2021) proposed a domain adaptation method using neuron pruning to target the problem of catastrophic forgetting of the general domain when fine-tuning a model for a target domain. They introduced a three step adaptation process: i) rank the most important neurons based on their importance, ii) prune the unimportant neurons from the network and retrain with student-teacher framework, iii) expand the network to its original size and fine-tune towards in-domain freezing the salient neurons and adjusting only the unimportant neurons.…”
Section: Domain Adaptationmentioning
confidence: 99%
“…Identifying the salient neurons with respect to a domain can be effectively used for domain adaptation and generalization. Gu et al (2021) proposed a domain adaptation method using neuron pruning to target the problem of catastrophic forgetting of the general domain when fine-tuning a model for a target domain. They introduced a three step adaptation process: i) rank the most important neurons based on their importance, ii) prune the unimportant neurons from the network and retrain with student-teacher framework, iii) expand the network to its original size and fine-tune towards in-domain freezing the salient neurons and adjusting only the unimportant neurons.…”
Section: Domain Adaptationmentioning
confidence: 99%
“…Parameters to be frozen do not necessarily have to come from the same subnetwork. More recent work finds sparse or underused areas of the network that can be easily adapted to new domains (Gu et al, 2021;Liang et al, 2020). A related idea is to factorize existing model components into general domain and domain-specific before tuning (Deng et al, 2020).…”
Section: Freezing Parametersmentioning
confidence: 99%
“…Besides, Lan et al (2020) presents two parameter reduction techniques to lower memory consumption and increase the training speed of BERT. Gu et al (2021) prune then expand the model neurons or parame-ters based on importance on domain adaptation of neural machine translation.…”
Section: Related Workmentioning
confidence: 99%