Proceedings of the 8th Workshop on Asian Translation (WAT2021) 2021
DOI: 10.18653/v1/2021.wat-1.11
|View full text |Cite
|
Sign up to set email alerts
|

Zero-pronoun Data Augmentation for Japanese-to-English Translation

Abstract: For Japanese-to-English translation, zero pronouns in Japanese pose a challenge, since the model needs to infer and produce the corresponding pronoun in the target side of the English sentence. However, although fully resolving zero pronouns often needs discourse context, in some cases, the local context within a sentence gives clues to the inference of the zero pronoun. In this study, we propose a data augmentation method that provides additional training signals for the translation model to learn correlation… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
5
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(6 citation statements)
references
References 17 publications
(11 reference statements)
1
5
0
Order By: Relevance
“…Language Bias. Most works used Chinese and Japanese datasets as testbed for training ZP models (Song et al, 2020;Ri et al, 2021). However, there were limited data available for other prodrop languages (e.g.…”
Section: Discussion and Findingsmentioning
confidence: 99%
See 4 more Smart Citations
“…Language Bias. Most works used Chinese and Japanese datasets as testbed for training ZP models (Song et al, 2020;Ri et al, 2021). However, there were limited data available for other prodrop languages (e.g.…”
Section: Discussion and Findingsmentioning
confidence: 99%
“…They trained an external model on the ZP data to recover the ZP information in the input sequence of the MT model (Tan et al, 2019;Ohtani et al, 2019;Tan et al, 2021) or correct the errors in the translation outputs (Voita et al, 2019). Others aimed to up-sample the training data for the ZPT task (Sugiyama and Yoshinaga, 2019;Kimura et al, 2019;Ri et al, 2021). They preferred to improve the ZPT performance via a data augmentation without modifying the MT architecture (Wang et al, 2016a;Sugiyama and Yoshinaga, 2019).…”
Section: Data-level Methods Do Not Change Modelmentioning
confidence: 99%
See 3 more Smart Citations