2008
DOI: 10.1109/icassp.2008.4518805
|View full text |Cite
|
Sign up to set email alerts
|

Extracting clues from human interpreter speech for spoken language translation

Abstract: In previous work, we reported dramatic improvements in automatic speech recognition (ASR) and spoken language translation (SLT) gained by applying information extracted from spoken human interpretations. These interpretations were artificially created by collecting read sentences from a clean parallel text corpus. Real human interpretations are significantly different. They suffer from frequent synopses, omissions and self-corrections. Expressing these differences in BLEU score by evaluating human interpretati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2009
2009
2017
2017

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 9 publications
(4 citation statements)
references
References 7 publications
0
4
0
Order By: Relevance
“…One can combine speech with a text stream, usually for an application such as machine-aided human translation [1,2], in which a human translator dictates the translation, rather than typing it. Also, a few works have looked into combining several speech streams [3,4], to improve ASR and MT systems in a simultaneous or consecutive interpretation scenario.…”
Section: Relation To Previous Workmentioning
confidence: 99%
“…One can combine speech with a text stream, usually for an application such as machine-aided human translation [1,2], in which a human translator dictates the translation, rather than typing it. Also, a few works have looked into combining several speech streams [3,4], to improve ASR and MT systems in a simultaneous or consecutive interpretation scenario.…”
Section: Relation To Previous Workmentioning
confidence: 99%
“…We used ASR hypotheses as well as reference transcripts for the experiments, whereas the Spanish hypotheses were generated with a system trained within TC-STAR on Parliament plenary sessions (Stüker et al 2007;Paulik and Waibel 2008). The case-insensitive WER was 8.4%.…”
Section: Translatable Speech Segmentsmentioning
confidence: 99%
“…Further, the hypotheses of both ASR systems can be tied together in a parallel training corpus suitable for TM training, as shown in [1]. Similar to previous works [2,3,4], we exploit the parallel information given in the respective other language audio stream to bias the ASR systems for an improved transcription performance. In the proposed context, such an improved ASR performance directly affects the quality of the extracted training data.…”
Section: System Architecturementioning
confidence: 99%