2016 Second International Conference on Event-Based Control, Communication, and Signal Processing (EBCCSP) 2016
DOI: 10.1109/ebccsp.2016.7605233
|View full text |Cite
|
Sign up to set email alerts
|

Steering a predator robot using a mixed frame/event-driven convolutional neural network

Abstract: Abstract-This paper describes the application of a Convolutional Neural Network (CNN) in the context of a predator/prey scenario. The CNN is trained and run on data from a Dynamic and Active Pixel Sensor (DAVIS) mounted on a Summit XL robot (the predator), which follows another one (the prey). The CNN is driven by both conventional image frames and dynamic vision sensor "frames" that consist of a constant number of DAVIS ON and OFF events. The network is thus "data driven" at a sample rate proportional to the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
67
0
1

Year Published

2019
2019
2020
2020

Publication Types

Select...
5
3
1

Relationship

2
7

Authors

Journals

citations
Cited by 96 publications
(71 citation statements)
references
References 7 publications
0
67
0
1
Order By: Relevance
“…Selecting the appropriate input representation of a set of events for a neural network is still a challenging problem. Prior works such as Moeys et al [14] and Maqueda et al [11] generate an event image by summing the number of events at each pixel. However, this discards the rich temporal information in the events, and is susceptible to motion blur.…”
Section: Input: the Discretized Event Volumementioning
confidence: 99%
“…Selecting the appropriate input representation of a set of events for a neural network is still a challenging problem. Prior works such as Moeys et al [14] and Maqueda et al [11] generate an event image by summing the number of events at each pixel. However, this discards the rich temporal information in the events, and is susceptible to motion blur.…”
Section: Input: the Discretized Event Volumementioning
confidence: 99%
“…RoshamboNet is a 5-layer, 20 MOp, 114k weight CNN architecture, described in Table V, trained to play the rockscissors-paper game [31]. This network can classify input images of size 64x64 obtained from the DVS of a DAVIS camera using the same training and feature extraction stage approach from [32]. The network outputs 4 classes: "rock", "scissors", "paper" or "background" from each feature vector.…”
Section: Roshambonetmentioning
confidence: 99%
“…The face detector is a small CNN designed to recognize whether a face is present or absent in an image obtained from the DAVIS camera. The DVS events are accumulated into 36x36 input images, again using the method of [32]. The network was trained on a dataset of 1800k frames collected from public face datasets and labeled DAVIS frames.…”
Section: Face Detector Cnnmentioning
confidence: 99%
“…Hence, we argue that researchers in the field should select a dataset onto which SNN accelerators could be compared fairly, where timing information is relevant, and no input conversion is required. Several event-driven datasets obtained with bioinspired image sensors have already been proposed [18,125,169,204].…”
Section: Low Power Spiking Machine Learningmentioning
confidence: 99%