2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2016
DOI: 10.1109/icassp.2016.7472829
|View full text |Cite
|
Sign up to set email alerts
|

CUED-RNNLM — An open-source toolkit for efficient training and evaluation of recurrent neural network language models

Abstract: In recent years, recurrent neural network language models (RNNLMs) have become increasingly popular for a range of applications including speech recognition. However, the training of RNNLMs is computationally expensive, which limits the quantity of data, and size of network, that can be used. In order to fully exploit the power of RNNLMs, efficient training implementations are required. This paper introduces an open-source toolkit, the CUED-RNNLM toolkit, which supports efficient GPU-based training of RNNLMs. … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
48
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
5
3
2

Relationship

1
9

Authors

Journals

citations
Cited by 73 publications
(48 citation statements)
references
References 23 publications
0
48
0
Order By: Relevance
“…In addition, a recurrent neural network language model (RNNLM) [29] was also used to refine the result of the first pass decoding. The CUED-RNNLM Toolkit v1.0 [30] was used to train the RNNLM 1 The splicing indexes per layer can be described as {-1,0,1} {-1,0,1} {-1,0,1,2} {-3,0,3} {-3,0,3} {-6,-3,0} {0} using the notation of [8,11]. 2 The architecture can be described as {-2,-1,0,1,2} {-1,0,1} L {-3,0,3} {-3,0,3} L {-3,0,3} {-3,0,3} L, where L represents an LSTMP layer with 512 cells and 128-dimensional recurrent and non-recurrent projections, using notation of [8,11].…”
Section: Methodsmentioning
confidence: 99%
“…In addition, a recurrent neural network language model (RNNLM) [29] was also used to refine the result of the first pass decoding. The CUED-RNNLM Toolkit v1.0 [30] was used to train the RNNLM 1 The splicing indexes per layer can be described as {-1,0,1} {-1,0,1} {-1,0,1,2} {-3,0,3} {-3,0,3} {-6,-3,0} {0} using the notation of [8,11]. 2 The architecture can be described as {-2,-1,0,1,2} {-1,0,1} L {-3,0,3} {-3,0,3} L {-3,0,3} {-3,0,3} L, where L represents an LSTMP layer with 512 cells and 128-dimensional recurrent and non-recurrent projections, using notation of [8,11].…”
Section: Methodsmentioning
confidence: 99%
“…Our RNN-LMs are trained and evaluated using the CUED-RNNLM toolkit [58]. Our RNN-LM configuration has several distinctive features, as described below.…”
Section: Rnn-lm Setupmentioning
confidence: 99%
“…Kneser-Ney smoothing is used for building the 4-gram LM using the corresponding options provided in SRILM. The RNNLMs were trained used a modified version of the CUED-RNNLM toolkit [50]. For training the baseline RNNLM, we used the full LM 1&LM 2 text, together with a 60k vocabulary for the input word list and a 50k vocabulary for the output word list.…”
Section: A Asr Results 1) Experimental Setupmentioning
confidence: 99%