2019
DOI: 10.1109/tsmc.2017.2787482
|View full text |Cite
|
Sign up to set email alerts
|

Continuous Visual World Modeling for Autonomous Robot Manipulation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0
1

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3

Relationship

3
4

Authors

Journals

citations
Cited by 17 publications
(10 citation statements)
references
References 35 publications
0
8
0
1
Order By: Relevance
“…Moreover, instead of extracting temporal features with HMMs and temporal pyramid pooling [34] to capture unexpected trends in the data for a fixed amount of time steps, we employ self-attention enabled LSTMs to capture anomaly indicators that may be observed long before the anomaly occurrences. In our previous work [6], we present a symbolic-level anomaly identification method that processes the outputs of a visual scene modeling system [35], proprioceptive sensors and auditory data to identify anomalies with preprocessed hand-crafted features. In this study, we extend it by presenting a three stream anomaly identification framework that extracts lowlevel features from 2D images directly without considering high-level symbolic domain symbols which does not require any hand-crafted feature engineering effort.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Moreover, instead of extracting temporal features with HMMs and temporal pyramid pooling [34] to capture unexpected trends in the data for a fixed amount of time steps, we employ self-attention enabled LSTMs to capture anomaly indicators that may be observed long before the anomaly occurrences. In our previous work [6], we present a symbolic-level anomaly identification method that processes the outputs of a visual scene modeling system [35], proprioceptive sensors and auditory data to identify anomalies with preprocessed hand-crafted features. In this study, we extend it by presenting a three stream anomaly identification framework that extracts lowlevel features from 2D images directly without considering high-level symbolic domain symbols which does not require any hand-crafted feature engineering effort.…”
Section: Literature Reviewmentioning
confidence: 99%
“…The templates of the objects are stored in the knowledge base (KB) of the robot and during the execution, point clouds from the scene are compared with the templates in the knowledge base of the robot. If a point cloud is matched with a template, the object is recognized with a similarity measure [37].…”
Section: Visionmentioning
confidence: 99%
“…Clusters that are larger or smaller than a predefined threshold are discarded. Note that the segments clustered with the segmentation algorithm are classified as unknown objects, these segments do not have an object type [37].…”
Section: Visionmentioning
confidence: 99%
See 2 more Smart Citations