Interspeech 2022 2022
DOI: 10.21437/interspeech.2022-10931
|View full text |Cite
|
Sign up to set email alerts
|

Qualitative Evaluation of Language Model Rescoring in Automatic Speech Recognition

Abstract: Evaluating automatic speech recognition (ASR) systems is a classical but difficult and still open problem, which often boils down to focusing only on the word error rate (WER). However, this metric suffers from many limitations and does not allow an in-depth analysis of automatic transcription errors. In this paper, we propose to study and understand the impact of rescoring using language models in ASR systems by means of several metrics often used in other natural language processing (NLP) tasks in addition t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 13 publications
(17 reference statements)
0
2
0
Order By: Relevance
“…Next, we examine three semantic metrics based on word embedding representations. The first one, Embedding Error Rate (EmbER) [2], is a WER where substitution errors are weighted according to the cosine distance between the reference and the substitute word embeddings obtained from fastText [14,3]. The second one, SemDist [20], involves calculating the cosine similarity between the reference and hypothesis using embeddings obtained at the sentence level.…”
Section: Evaluation Metricsmentioning
confidence: 99%
See 1 more Smart Citation
“…Next, we examine three semantic metrics based on word embedding representations. The first one, Embedding Error Rate (EmbER) [2], is a WER where substitution errors are weighted according to the cosine distance between the reference and the substitute word embeddings obtained from fastText [14,3]. The second one, SemDist [20], involves calculating the cosine similarity between the reference and hypothesis using embeddings obtained at the sentence level.…”
Section: Evaluation Metricsmentioning
confidence: 99%
“…Consequently, there has been a growing interest in developing new metrics to evaluate ASR systems. Some researchers [23,27,13,20,2] have therefore started exploring alternative metrics that can more accurately assess the quality and effectiveness of automatic transcriptions. Similarly, these issues have been observed in the field of machine translation.…”
Section: Introductionmentioning
confidence: 99%