Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2022
DOI: 10.18653/v1/2022.naacl-main.186
|View full text |Cite
|
Sign up to set email alerts
|

Training Mixed-Domain Translation Models via Federated Learning

Abstract: Training mixed-domain translation models is a complex task that demands tailored architectures and costly data preparation techniques. In this work, we leverage federated learning (FL) in order to tackle the problem. Our investigation demonstrates that with slight modifications in the training process, neural machine translation (NMT) engines can be easily adapted when an FL-based aggregation is applied to fuse different domains. Experimental results also show that engines built via FL are able to perform on p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
0
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 1 publication
(1 reference statement)
0
0
0
Order By: Relevance
“…Federated Learning in NLP In order to protect data privacy, federated Learning (FL) has attracted a lot of attention from both academic and industrial (Konečnỳ et al, 2016;McMahan et al, 2017;Yang et al, 2019;Kairouz et al, 2021). With more and more language assistance products being applied in real-world applications, FL has also increasingly appeared in the community of NLP to address the problem of privacy leakage, such as machine translation (Passban et al, 2022;Du et al, 2023) and question answering Ait-Mlouk et al, 2022), and so on Cai et al, 2023;.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Federated Learning in NLP In order to protect data privacy, federated Learning (FL) has attracted a lot of attention from both academic and industrial (Konečnỳ et al, 2016;McMahan et al, 2017;Yang et al, 2019;Kairouz et al, 2021). With more and more language assistance products being applied in real-world applications, FL has also increasingly appeared in the community of NLP to address the problem of privacy leakage, such as machine translation (Passban et al, 2022;Du et al, 2023) and question answering Ait-Mlouk et al, 2022), and so on Cai et al, 2023;.…”
Section: Related Workmentioning
confidence: 99%
“…(iii) Efficient communication. For example, Passban et al (2022) presents a dynamic pulling FL method to dynamically control the communication bandwidth. Du et al (2023) presents a federated nearest neighbor framework to reduce the communication overhead.…”
Section: Related Workmentioning
confidence: 99%
“…Language modeling is one of the fundamental tasks in Natural Language Processing and FL for language modeling recently attracted attention in academia and industry [29,37,6].…”
Section: Related Workmentioning
confidence: 99%