2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2017
DOI: 10.1109/icassp.2017.7953263
|View full text |Cite
|
Sign up to set email alerts
|

Recurrent neural network language models for keyword search

Abstract: Recurrent neural network language models (RNNLMs) have becoming increasingly popular in many applications such as automatic speech recognition (ASR). Significant performance improvements in both perplexity and word error rate over standard n-gram LMs have been widely reported on ASR tasks. In contrast, published research on using RNNLMs for keyword search systems has been relatively limited. In this paper the application of RNNLMs for the IARPA Babel keyword search task is investigated. In order to supplement … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2017
2017
2019
2019

Publication Types

Select...
2
2
1

Relationship

1
4

Authors

Journals

citations
Cited by 8 publications
(9 citation statements)
references
References 20 publications
(24 reference statements)
0
9
0
Order By: Relevance
“…Lattice rescoring is impractical for bi-RNNLMs as the word probability calculations require information from the complete sentence. However, lattices are very important in a range of downstream applications, including confidence score estimation [21], keyword search [22] and confusion network decoding [23].…”
Section: Bi-directional Rnnlmsmentioning
confidence: 99%
See 1 more Smart Citation
“…Lattice rescoring is impractical for bi-RNNLMs as the word probability calculations require information from the complete sentence. However, lattices are very important in a range of downstream applications, including confidence score estimation [21], keyword search [22] and confusion network decoding [23].…”
Section: Bi-directional Rnnlmsmentioning
confidence: 99%
“…In these experiments the performance of su-RNNLMs, which can be directly applied to lattices is compared to uni-RNNLMs. In [41] uni-RNNLMs were demonstrated to be effective for KWS. A total about 50 hours of transcribed conversational telephone speech data are provided to build the ASR and keyword search systems.…”
Section: Experiments On Abstractearchmentioning
confidence: 99%
“…On the other hand, the subword-based approach has the unique advantage that it can detect terms that consist of words that are not in the vocabulary of the recognizer, i.e., out-ofvocabulary (OOV) terms. The combination of these two approaches has been proposed in order to exploit the relative advantages of word and subword-based strategies [17,32,33,36,44,[63][64][65][66][67][68][69][70].…”
Section: Spoken Term Detection Overviewmentioning
confidence: 99%
“…It is not practical for bi-RNNLMs to be used for lattice rescoring and generation as both the complete previous and future context information are required. However, lattices are very useful in many applications, such as confidence score estimation [9], keyword search [10] and confusion network decoding [11]. In contrast, su-RNNLMs require a fixed number of succeeding words, instead of the complete future context information.…”
Section: Lattice Rescoringmentioning
confidence: 99%
“…However, the ability to manipulate lattices is very important in many speech applications. Lattices can be used for a wide range of downstream applications, such as confidence score estimation [9], keyword search [10] and confusion network decoding [11]. In order to address these issues, a novel model structure, succeeding word RNNLMs (su-RNNLMs), is proposed in this paper.…”
Section: Introductionmentioning
confidence: 99%