2007 IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP '07 2007
DOI: 10.1109/icassp.2007.366923
|View full text |Cite
|
Sign up to set email alerts
|

The AMI System for the Transcription of Speech in Meetings

Abstract: This paper describes the AMI transcription system for speech in meetings developed in collaboration by five research groups. The system includes generic techniques such as discriminative and speaker adaptive training, vocal tract length normalisation, heteroscedastic linear discriminant analysis, maximum likelihood linear regression, and phone posterior based features, as well as techniques specifically designed for meeting data. These include segmentation and cross-talk suppression, beam-forming, domain adapt… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
68
0

Year Published

2007
2007
2018
2018

Publication Types

Select...
5
3
2

Relationship

2
8

Authors

Journals

citations
Cited by 91 publications
(68 citation statements)
references
References 8 publications
(16 reference statements)
0
68
0
Order By: Relevance
“…In the following we briefly address issues of the domain, followed by a brief description of the essential components of a meeting transcription system and the performance in recent NIST evaluations. For a more elaborate description of the systems, the interested reader is referred to [14].…”
Section: Meeting Speech Recognitionmentioning
confidence: 99%
“…In the following we briefly address issues of the domain, followed by a brief description of the essential components of a meeting transcription system and the performance in recent NIST evaluations. For a more elaborate description of the systems, the interested reader is referred to [14].…”
Section: Meeting Speech Recognitionmentioning
confidence: 99%
“…Acoustic segmentation and speech/non-speech detection remains an important problem, with nearly 10% of errors in our current system resulting from errors in the speech/non-speech detection component. A feature of the systems developed for meeting recognition is the use of multiple recognition passes, cross-adaptation and model combination (Hain et al 2007). In particular successive passes make use of more detailed-and more diverse-acoustic and language models.…”
Section: Long-context Featuresmentioning
confidence: 99%
“…Automatic transcriptions of the AMI meeting corpus were obtained using the AMI-ASR system [13]. This LVCSR system is based on decision tree clustered crossword triphone hidden Markov models, and a trigram language model.…”
Section: ) Speech Recognitionmentioning
confidence: 99%