2018
DOI: 10.1109/lra.2018.2860057
|View full text |Cite
|
Sign up to set email alerts
|

Deep Episodic Memory: Encoding, Recalling, and Predicting Episodic Experiences for Robot Action Execution

Abstract: We present a novel deep neural network architecture for representing robot experiences in an episodic-like memory which facilitates encoding, recalling, and predicting action experiences. Our proposed unsupervised deep episodic memory model 1) encodes observed actions in a latent vector space and, based on this latent encoding, 2) infers most similar episodes previously experienced, 3) reconstructs original episodes, and 4) predicts future frames in an end-to-end fashion. Results show that conceptually similar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0
1

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 32 publications
(13 citation statements)
references
References 19 publications
0
10
0
1
Order By: Relevance
“…The majority of work in behaviour cloning operates on a set of configuration-space trajectories that can be collected via tele-operation [15], [16], kinesthetic teaching [17], [18], sensors on a human demonstrator [19], [20], [21], [22], through motion planners [5], or even by observing humans directly. Expanding further on the latter, learning by observing humans has previously been achieved through hand-designed mappings between human actions and robot actions [1], [2], [23], visual activity recognition and explicit handtracking [24], [25], and more recently by a system that infers actions from a single video of a human via an end-toend trained system [4].…”
Section: Related Workmentioning
confidence: 99%
“…The majority of work in behaviour cloning operates on a set of configuration-space trajectories that can be collected via tele-operation [15], [16], kinesthetic teaching [17], [18], sensors on a human demonstrator [19], [20], [21], [22], through motion planners [5], or even by observing humans directly. Expanding further on the latter, learning by observing humans has previously been achieved through hand-designed mappings between human actions and robot actions [1], [2], [23], visual activity recognition and explicit handtracking [24], [25], and more recently by a system that infers actions from a single video of a human via an end-toend trained system [4].…”
Section: Related Workmentioning
confidence: 99%
“…This is not a simple question as decisions must be made about which students, if any, to prioritize: weaker students who require help to meet minimum learning requirements, or stronger students who are trying hard to learn. Additionally, another attractive prospect would be if a robot can itself learn to improve its knowledge base, e.g., from YouTube videos [75]; domain-specific learning has been demonstrated in various studies but general learning across various topics, and machine creativity for generating new content, are still desirable goals for future work.…”
Section: Future Workmentioning
confidence: 99%
“…We focus on this problem of learning from one video demonstration of a human performing a task, in combination with human and robot demonstration data collected on other tasks. Prior work has proposed to resolve the correspondence problem by hand, for example, by manually specifying how human grasp poses correspond to robot grasps [20] or by manually defining how human activities or commands translate into robot actions [58,23,30,37,40]. By utilizing demonstration data of how humans and robots perform each task, our approach learns the correspondence between the human and robot implicitly.…”
Section: Introductionmentioning
confidence: 99%