Proceedings of the Fourth Workshop on Neural Generation and Translation 2020
DOI: 10.18653/v1/2020.ngt-1.7
|View full text |Cite
|
Sign up to set email alerts
|

A Deep Reinforced Model for Zero-Shot Cross-Lingual Summarization with Bilingual Semantic Similarity Rewards

Abstract: Cross-lingual text summarization aims at generating a document summary in one language given input in another language. It is a practically important but under-explored task, primarily due to the dearth of available data. Existing methods resort to machine translation to synthesize training data, but such pipeline approaches suffer from error propagation. In this work, we propose an end-to-end cross-lingual text summarization model. The model uses reinforcement learning to directly optimize a bilingual semanti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
11
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 9 publications
(11 citation statements)
references
References 19 publications
0
11
0
Order By: Relevance
“…Conventional cross-lingual summarization methods mainly focus on incorporating bilingual information into the pipeline methods (Leuski et al, 2003;Ouyang et al, 2019;Orȃsan and Chiorean, 2008;Wan et al, 2010;Wan, 2011;Yao et al, 2015;Zhang et al, 2016b), i.e., translation and then summarization or summarization and then translation. Due to the difficulty of acquiring cross-lingual summarization dataset, some previous researches focus on constructing datasets (Ladhak et al, 2020;Scialom et al, 2020;Yela-Bello et al, 2021;Zhu et al, 2019;Hasan et al, 2021;Perez-Beltrachini and Lapata, 2021;Varab and Schluter, 2021), mixed-lingual pre-training (Xu et al, 2020), knowledge distillation (Nguyen and Tuan, 2021), contrastive learning (Wang et al, 2021) or zero-shot approaches (Ayana et al, 2018;Duan et al, 2019;Dou et al, 2020), i.e., using machine translation (MT) or monolingual summarization (MS) or both to train the CLS system. Among them, Zhu et al (2019) propose to use roundtrip translation strategy to obtain large-scale CLS datasets and then present two multi-task learning methods for CLS.…”
Section: Related Workmentioning
confidence: 99%
“…Conventional cross-lingual summarization methods mainly focus on incorporating bilingual information into the pipeline methods (Leuski et al, 2003;Ouyang et al, 2019;Orȃsan and Chiorean, 2008;Wan et al, 2010;Wan, 2011;Yao et al, 2015;Zhang et al, 2016b), i.e., translation and then summarization or summarization and then translation. Due to the difficulty of acquiring cross-lingual summarization dataset, some previous researches focus on constructing datasets (Ladhak et al, 2020;Scialom et al, 2020;Yela-Bello et al, 2021;Zhu et al, 2019;Hasan et al, 2021;Perez-Beltrachini and Lapata, 2021;Varab and Schluter, 2021), mixed-lingual pre-training (Xu et al, 2020), knowledge distillation (Nguyen and Tuan, 2021), contrastive learning (Wang et al, 2021) or zero-shot approaches (Ayana et al, 2018;Duan et al, 2019;Dou et al, 2020), i.e., using machine translation (MT) or monolingual summarization (MS) or both to train the CLS system. Among them, Zhu et al (2019) propose to use roundtrip translation strategy to obtain large-scale CLS datasets and then present two multi-task learning methods for CLS.…”
Section: Related Workmentioning
confidence: 99%
“…To solve this problem, propose a mix-lingual XLS model which is pre-trained with MLM, DAE, MS, TSC and MT tasks 17 . Dou et al (2020) utilize the XLS, MT and MS to pretrain the XLS model. Wang et al (2022b) focus on dialogue-oriented XLS and extend mBART-50 with AcI, UP, MS and MT tasks via the second pretraining stage.…”
Section: Pre-training Frameworkmentioning
confidence: 99%
“…Wang et al (2022b) focus on dialogue-oriented XLS and extend mBART-50 with AcI, UP, MS and MT tasks via the second pretraining stage. Note that , Dou et al (2020) and Wang et al (2022b) only focus on XLS task. Furthermore, ∆LM (Ma et al, 2021) and mT6 (Chi et al, 2021a) are presented towards general cross-lingual abilities.…”
Section: Pre-training Frameworkmentioning
confidence: 99%
See 1 more Smart Citation
“…With the widespread use of deep neural networks, many Transformer-based end-to-end methods [4], [5], [8], [9], [36], [48], [49] are proposed to directly understand bilingual semantics and avoid the error propagation problem. Due to lacking large-scale CLS datasets, Duan et al [9] and Dou et al [8] explore training end-to-end models with zeroshot learning.…”
Section: Introductionmentioning
confidence: 99%