Interspeech 2017 2017
DOI: 10.21437/interspeech.2017-1320
|View full text |Cite
|
Sign up to set email alerts
|

NMT-Based Segmentation and Punctuation Insertion for Real-Time Spoken Language Translation

Abstract: Insertion of proper segmentation and punctuation into an ASR transcript is crucial not only for the performance of subsequent applications but also for the readability of the text. In a simultaneous spoken language translation system, the segmentation model has to fulfill real-time constraints and minimize latency as well. In this paper, we show the successful integration of an attentional encoder-decoder-based segmentation and punctuation insertion model into a real-time spoken language translation system. Th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
42
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 41 publications
(44 citation statements)
references
References 13 publications
0
42
0
Order By: Relevance
“…The ones with plain text outperform the ones with To explore the impact of min_words_cut value to the quality of the result, we performed the experiment on sequenceto-sequence LSTM model with the overlapping of 15 words and min_words_cut ranges from 0 to 15. The outcome shown in Figure 5 indicates that f1-scores peak in the middle range of chunk size (4)(5)(6)(7)(8)(9)(10). It demonstrate that predictions of uppercase and lowercase are stable and independent from min_words_cut.…”
Section: Evaluation On Plain-text Model and Encoded-text Modelmentioning
confidence: 83%
See 2 more Smart Citations
“…The ones with plain text outperform the ones with To explore the impact of min_words_cut value to the quality of the result, we performed the experiment on sequenceto-sequence LSTM model with the overlapping of 15 words and min_words_cut ranges from 0 to 15. The outcome shown in Figure 5 indicates that f1-scores peak in the middle range of chunk size (4)(5)(6)(7)(8)(9)(10). It demonstrate that predictions of uppercase and lowercase are stable and independent from min_words_cut.…”
Section: Evaluation On Plain-text Model and Encoded-text Modelmentioning
confidence: 83%
“…An example is shown in Figure 4. We prepared 2 formats of training data: plain text and encoded text [9]. Both formats takes the lowercase text without punctuation as input.…”
Section: Data Preparationmentioning
confidence: 99%
See 1 more Smart Citation
“…Previous work studied different decoding strategies to reduce the latency for real-time spoken language processing systems, including overlapping windows [7], streaming input scheme [8], and overlapped-chunk split and merging strategy [9]. However, the input text for inference in these decoding strategies does not always begin with the first word of a sentence.…”
Section: Introductionmentioning
confidence: 99%
“…Punctuation is essential for grammaticality, readability, and (in the case of a number of different tasks), subsequent processing. Thus, correct sentence segmentation and punctuation of recognized speech improves the quality of machine translation [6,7,24,26], and missing periods and commas in machine generated text results in suboptimal information extraction from speech [13,15]. Also, most of the data-driven parsing models use punctuation as features.…”
Section: Introductionmentioning
confidence: 99%