2016 IEEE Intl Conference on Computational Science and Engineering (CSE) and IEEE Intl Conference on Embedded and Ubiquitous Co 2016
DOI: 10.1109/cse-euc-dcabes.2016.229
|View full text |Cite
|
Sign up to set email alerts
|

Echo State Networks-Based Reservoir Computing for MNIST Handwritten Digits Recognition

Abstract: Reservoir Computing is an attractive paradigm of recurrent neural network architecture, due to the ease of training and existing neuromorphic implementations. Successively applied on speech recognition and time series forecasting, few works have so far studied the behavior of such networks on computer vision tasks. Therefore we decided to investigate the ability of Echo State Networks to classify the digits of the MNIST database. We show that even if ESNs are not able to outperform state-of-the-art convolution… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
50
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 56 publications
(50 citation statements)
references
References 13 publications
0
50
0
Order By: Relevance
“…It works particularly well for analyzing time series input data due to its short-term memory [15] and high-dimensional encoding of the input [35,36]. The input images are hence converted into a "time series" by feeding the reservoir a column of the input image at each time point (as in [37]). The method of "temporalization" of the input (row-wise, columnwise, etc.)…”
Section: Methodsmentioning
confidence: 99%
“…It works particularly well for analyzing time series input data due to its short-term memory [15] and high-dimensional encoding of the input [35,36]. The input images are hence converted into a "time series" by feeding the reservoir a column of the input image at each time point (as in [37]). The method of "temporalization" of the input (row-wise, columnwise, etc.)…”
Section: Methodsmentioning
confidence: 99%
“…The third approach, inspired by [36], uses an explicit temporal encoding of the spatial visual information from the MNIST images in order to activate the recurrent dynamics of the reservoir computer. The idea here is to split the full image into smaller portions and feed them sequentially into the classifier.…”
Section: Column-wise Recurrent Modementioning
confidence: 99%
“…The separation can be done in various ways: columns or rows (overlapping or adjacent), sliding windows, sub-images, among others. Similarly to [36], we consider nonoverlapping columns, which transforms each image into 28 inputs of 28 dimensions. Therefore, such temporal encoding reduces the input dimensionality, but increases the processing time proportionally.…”
Section: Column-wise Recurrent Modementioning
confidence: 99%
“…This function has to be learned by the model in order to solve the task. It would have been possible to use the MNIST database (Schaetti, Salomon, & Couturier, 2016) instead of a regular font but this would have also complexified the task, and make the training period much longer, because a digit is processed only when a trigger is present. If we consider for example a sequence of 25,000 digits and a trigger probability of 0.01, this represents (in average) 250 triggers for the whole sequence and consequently only 25 presentations per digit.…”
Section: The Digit 1-value 1-gate Working Memory Taskmentioning
confidence: 99%