Interspeech 2007 2007
DOI: 10.21437/interspeech.2007-655
|View full text |Cite
|
Sign up to set email alerts
|

Optimizing sentence segmentation for spoken language translation

Abstract: The conventional approach in text-based machine translation (MT) is to translate complete sentences, which are conveniently indicated by sentence boundary markers. However, since such boundary markers are not available for speech, new methods are required that define an optimal unit for translation. Our experimental results show that with a segment length optimized for a particular MT system, intrasentence segmentation can improve translation performance (measured in BLEU) by up to 11% for Arabic Broadcast Con… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2007
2007
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 8 publications
0
4
0
Order By: Relevance
“…While following a cascaded approach, one can not directly chain modules such as ASR, MT, and TTS as it is a well known fact that spoken language has various idiosyncrasies. These include lack of well-formed sentences and disfluencies (Rao et al, 2007). Traditional machine translation systems are trained on well formed, written, and grammatical pairs of sentences.…”
Section: Approachesmentioning
confidence: 99%
“…While following a cascaded approach, one can not directly chain modules such as ASR, MT, and TTS as it is a well known fact that spoken language has various idiosyncrasies. These include lack of well-formed sentences and disfluencies (Rao et al, 2007). Traditional machine translation systems are trained on well formed, written, and grammatical pairs of sentences.…”
Section: Approachesmentioning
confidence: 99%
“…While following a cascaded approach, one can not directly chain modules such as ASR, MT, and TTS as it is a well known fact that spoken language has various idiosyncrasies. These include lack of well-formed sentences and disfluencies (Rao et al, 2007). Traditional machine translation systems are trained on well formed, written, and grammatical pairs of sentences.…”
Section: Approachesmentioning
confidence: 99%
“…Segmentation in SLT has been studied quite extensively in high-resource settings. Early work used kernel-based SVM models to predict sentence boundaries using language model probabilities along with prosodic features such as pause duration (Matusov et al, 2007;Rao et al, 2007) and part-of-speech features derived from a fixed window size (Rangarajan Sridhar et al, 2013). Other work has modeled the problem using hidden markov models (Shriberg et al, 2000;Gotoh and Renals, 2000;Christensen et al, 2001;Kim and Woodland, 2001) and conditional random fields (Liu et al, 2005;Lu and Ng, 2010).…”
Section: Related Workmentioning
confidence: 99%
“…While prior work has trained intermediate components to segment ASR output into sentence-like units (Matusov et al, 2007;Rao et al, 2007), these have primarily focused on highly resourced language pairs such as Arabic and Chinese. When the source language is low-resource, suitable training data may be very limited for ASR and MT, and even nonexistent for segmentation.…”
Section: Introductionmentioning
confidence: 99%