2012
DOI: 10.48550/arxiv.1211.5590
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Theano: new features and speed improvements

Abstract: Theano is a linear algebra compiler that optimizes a user's symbolically-specified mathematical computations to produce efficient low-level implementations. In this paper, we present new features and efficiency improvements to Theano, and benchmarks demonstrating Theano's performance relative to Torch7, a recently introduced machine learning library, and to RNNLM, a C++ library targeted at recurrent neural networks.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
105
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 81 publications
(105 citation statements)
references
References 7 publications
0
105
0
Order By: Relevance
“…The best model during the entire run was kept. All experiments were carried out using Python and implemented in Theano [16,17].…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…The best model during the entire run was kept. All experiments were carried out using Python and implemented in Theano [16,17].…”
Section: Resultsmentioning
confidence: 99%
“…More complex architectures are possible in our implementation, but are not considered here. The model was implemented in Python and Theano [16,17].…”
Section: Implementation Of the Modelmentioning
confidence: 99%
“…To calculate the momentum at a position in posterior space we need to calculate the derivative of the posterior density for the local environment. With the advances in automatic differentiation in tools like, for example, Theano [48], which is used in the pymc3 framework [49], it is possible to quickly calculate the derivatives of more complex functions.…”
Section: Sampling Algorithmmentioning
confidence: 99%
“…Various platforms facilitating the construction and evaluation of deep neural networks exist. They include Tensorflow [1], PyTorch [55], Theano [2], and Caffe [22]. Using these platforms, derivatives used in gradients are computed automatically and hence the user can concentrate on the design and optimization of the network architecture and other aspects of learning.…”
Section: Deep Learningmentioning
confidence: 99%