Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2018
DOI: 10.18653/v1/p18-1007
|View full text |Cite
|
Sign up to set email alerts
|

Subword Regularization: Improving Neural Network Translation Models with Multiple Subword Candidates

Abstract: Subword units are an effective way to alleviate the open vocabulary problems in neural machine translation (NMT). While sentences are usually converted into unique subword sequences, subword segmentation is potentially ambiguous and multiple segmentations are possible even with the same vocabulary. The question addressed in this paper is whether it is possible to harness the segmentation ambiguity as a noise to improve the robustness of NMT. We present a simple regularization method, subword regularization, wh… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

5
677
2
5

Year Published

2018
2018
2020
2020

Publication Types

Select...
5
3
1
1

Relationship

0
10

Authors

Journals

citations
Cited by 742 publications
(689 citation statements)
references
References 28 publications
(42 reference statements)
5
677
2
5
Order By: Relevance
“…A SentencePiece tokenizer [15] is also provided by the library. Subword tokenization [16] [17], such as that provided by SentencePiece, has been used in many recent NLP breakthroughs [18] [19].…”
Section: Textmentioning
confidence: 99%
“…A SentencePiece tokenizer [15] is also provided by the library. Subword tokenization [16] [17], such as that provided by SentencePiece, has been used in many recent NLP breakthroughs [18] [19].…”
Section: Textmentioning
confidence: 99%
“…For CTC training, we use word-pieces as our target. During training, the reference is tokenized to 5000 sub-word units using sentencepiece 1 with a uni-gram language model [15]. Neural networks are thus used to produce a posterior distribution for 5001 symbols (5000 sub-word units plus blank symbol) every frame.…”
Section: Target Unitsmentioning
confidence: 99%
“…We apply the word-piece approach of Kudo (2018), which computes a word-piece unigram LM using a word-piece inventory V P . Each wordpiece x i ∈ V P is associated with a unigram probability p(x i ).…”
Section: Morphologically Rich Languagesmentioning
confidence: 99%