Proceedings of the Workshop on Human Language Technology - HLT '94 1994
DOI: 10.3115/1075812.1075824
|View full text |Cite
|
Sign up to set email alerts
|

1993 benchmark tests for the ARPA spoken language program

Abstract: This paper reports results obtained in benchmark tests conducted within the ARPA Spoken Language program in November and December of 1993. In addition to ARPA contractors, participants included a number of %olunteers", including foreign participants from Canada, France, Germany, and the United Kingdom. The body of the paper is limited to an outline of the structure of the tests and presents highlights and discussion of selected results. Detailed tabulations of reported "official" results, and additional explan… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

6
61
0

Year Published

1994
1994
2013
2013

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 95 publications
(67 citation statements)
references
References 16 publications
6
61
0
Order By: Relevance
“…These 83 transformation matrices were used to build bases for the EMLLR and the PARAFAC2-based model. For the adaptation test, we used the adaptation data of 8 testing speakers from the WSJ0 corpus, i.e., the November 92 NIST evaluation set [11]. We used 1 to 5 utterances from the adaptation set (an adaptation utterance was about 6 s in length).…”
Section: Methodsmentioning
confidence: 99%
“…These 83 transformation matrices were used to build bases for the EMLLR and the PARAFAC2-based model. For the adaptation test, we used the adaptation data of 8 testing speakers from the WSJ0 corpus, i.e., the November 92 NIST evaluation set [11]. We used 1 to 5 utterances from the adaptation set (an adaptation utterance was about 6 s in length).…”
Section: Methodsmentioning
confidence: 99%
“…After working on isolated-word, speaker-dependent systems for many years, since 1992 the community has moved towards very-largevocabulary (20,000 words and more), high-perplexity ( ¡ ¥ ¤ § ¦ ) , speaker-independent, continuous speech recognition. The best system in 1994 achieved an error rate of 7.2% on read sentences drawn from North American business news (Pallett, Fiscus, et al, 1994).…”
Section: ¡ ¡ £ ¢mentioning
confidence: 99%
“…In the ARPA program, the air travel planning domain has been chosen to support evaluation of spoken language systems (Pallett, 1991;Pallett, 1992;Pallett, Dahlgren, et al, 1992;Pallett, Fisher, et al, 1990;Pallett, Fiscus, et al, 1993;Pallett, Fiscus, et al, 1994;Pallett, Fiscus, et al, 1995). Vocabularies for these systems are usually about 2000 words.…”
Section: State Of the Artmentioning
confidence: 99%
“…The training set consists of 60 hours of speech and the so-called WSJ test consists of 215 utterances [16]. The WSJ test is a 5k Hub test set.…”
Section: Wall Street Journalmentioning
confidence: 99%