Interspeech 2010 2010
DOI: 10.21437/interspeech.2010-343
|View full text |Cite
|
Sign up to set email alerts
|

Recurrent neural network based language model

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
466
0
8

Year Published

2016
2016
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 3,037 publications
(474 citation statements)
references
References 9 publications
0
466
0
8
Order By: Relevance
“…We conduct hyperparameter search, model introspection, and ablation studies on the English Penn Treebank (PTB) (Marcus, Santorini, and Marcinkiewicz 1993), utilizing the standard training (0-20), validation (21-22), and test (23-24) splits along with pre-processing by Mikolov et al (2010). With approximately 1m tokens and |V| = 10k, this version has been extensively used by the language modeling community and is publicly available.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…We conduct hyperparameter search, model introspection, and ablation studies on the English Penn Treebank (PTB) (Marcus, Santorini, and Marcinkiewicz 1993), utilizing the standard training (0-20), validation (21-22), and test (23-24) splits along with pre-processing by Mikolov et al (2010). With approximately 1m tokens and |V| = 10k, this version has been extensively used by the language modeling community and is publicly available.…”
Section: Methodsmentioning
confidence: 99%
“…Neural Language Models (NLM) encompass a rich family of neural network architectures for language modeling. Some example architectures include feed-forward (Bengio, Ducharme, and Vincent 2003), recurrent (Mikolov et al 2010), sum-product (Cheng et al 2014), log-bilinear (Mnih and Hinton 2007), and convolutional (Wang et al 2015) networks.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…We use Penn Treebank Dataset (henceforth PTB) (Taylor, Marcus, and Santorini 2003) with pre-processing in (Mikolov et al 2010) and the War and Peace Dataset (henceforth WP) as the standard benchmarks for character-level language modeling. PTB contains a set of collected 2499 stories designed to allow the extraction of simple predicate and argument structure.…”
Section: Datasetsmentioning
confidence: 99%
“…We use the lexical semantics model and implementation, created by Jansen, Surdeanu, and Clark (2014), to generate domain-appropriate embeddings for a corpus of elementary science text. The embeddings are learned using the recurrent neural network language model (RNNLM) (Mikolov et al 2010;. Like any language model, a RNNLM estimates the probability of observing a word given the preceding context, but, in this process, it also learns word embeddings into a latent, conceptual space with a fixed number of dimensions.…”
Section: The Pointwise Mutual Information (Pmi) Solvermentioning
confidence: 99%