2020
DOI: 10.3390/electronics9101562
|View full text |Cite
|
Sign up to set email alerts
|

Decoding Strategies for Improving Low-Resource Machine Translation

Abstract: Pre-processing and post-processing are significant aspects of natural language processing (NLP) application software. Pre-processing in neural machine translation (NMT) includes subword tokenization to alleviate the problem of unknown words, parallel corpus filtering that only filters data suitable for training, and data augmentation to ensure that the corpus contains sufficient content. Post-processing includes automatic post editing and the application of various strategies during decoding in the translation… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
18
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3

Relationship

3
5

Authors

Journals

citations
Cited by 22 publications
(18 citation statements)
references
References 17 publications
0
18
0
Order By: Relevance
“…In particular, the number of training parameters for the APE model, in which d bal is set to 256, is lower than 5% of all the model parameters. This demonstrates that efficient training can be obtained via the BAL structure, and we can obtain an advantage with respect to the GPU usage or training time, while considering industrial services [11,30].…”
Section: E Question 2: Does the Adapter Structure Provide An Advantage In Ape?mentioning
confidence: 99%
“…In particular, the number of training parameters for the APE model, in which d bal is set to 256, is lower than 5% of all the model parameters. This demonstrates that efficient training can be obtained via the BAL structure, and we can obtain an advantage with respect to the GPU usage or training time, while considering industrial services [11,30].…”
Section: E Question 2: Does the Adapter Structure Provide An Advantage In Ape?mentioning
confidence: 99%
“…It is difficult to establish a sufficient hardware environment to provide services, except for large companies such as Google and Facebook. In other words, as training a model involves many parameters and a large amount of data, companies that do not have sufficient server or GPU environments will find it difficult to configure the service environment and improve performance using the latest model (Park et al, 2020c). Therefore, it is important to ensure that companies with insufficient environments can provide services while performing well against LRLs.…”
Section: Introductionmentioning
confidence: 99%
“…To solve this problem, many researches are being conducted on the way of improving the performance of NLP application software without changing the model through data pre and post-processing, typically in machine translation (Pal et al, 2016;Currey et al, 2017;Banerjee and Bhattacharyya, 2018;Koehn et al, 2018;Kudo, 2018;Park et al, 2020b). Reflecting this trend, we conducted a study on an optimized tokenization that can improve the performance of neural machine translation (NMT) without changing the model.…”
Section: Introductionmentioning
confidence: 99%