Our system is currently under heavy load due to increased usage. We're actively working on upgrades to improve performance. Thank you for your patience.
2017 IEEE International Conference on Computer Vision Workshops (ICCVW) 2017
DOI: 10.1109/iccvw.2017.279
|View full text |Cite
|
Sign up to set email alerts
|

An Object is Worth Six Thousand Pictures: The Egocentric, Manual, Multi-image (EMMI) Dataset

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(7 citation statements)
references
References 16 publications
0
7
0
Order By: Relevance
“…In (Abebe & Cavallaro, 2017c), stacked spectrograms of motion patterns extracted from and GTEA Gaze+ 8 . However, since the publication of (Nguyen et al, 2016) other datasets have appeared such as some datasets for manipulated object detection such as the EMMI dataset by Wang et al (2017a). In their paper 735 they also explore other manipulated object recognition datasets, which could be interesting to the reader.…”
Section: From Wearable or First-person Visionmentioning
confidence: 99%
“…In (Abebe & Cavallaro, 2017c), stacked spectrograms of motion patterns extracted from and GTEA Gaze+ 8 . However, since the publication of (Nguyen et al, 2016) other datasets have appeared such as some datasets for manipulated object detection such as the EMMI dataset by Wang et al (2017a). In their paper 735 they also explore other manipulated object recognition datasets, which could be interesting to the reader.…”
Section: From Wearable or First-person Visionmentioning
confidence: 99%
“…The first dataset is T-LESS [30], which contains 30 objects with no relevant texture. Followed by ToyBox [31] a dataset depicting 360 objects manipulated by a person. Toybox allowed us to evaluate how well the model scales (we selected ToyBox over the iCubWorld dataset [37] since, at the submission time, the latter contains only 28 instances).…”
Section: Methodsmentioning
confidence: 99%
“…In T-LESS [30], the model has to recognize unseen viewpoints. In Toybox [31] we used the hodgepodge videos for training and translations across the three axes for testing. We use ARC as in [4].…”
Section: B Triplet Lossmentioning
confidence: 99%
“…As a first example, data sets cover the sensation during human-environment interaction by measuring (mostly adult) humans directly during performing specific tasks, such as the KIT Motion-Language set for descriptions of whole-body poses (Plappert et al, 2016 ), the Multimodal-HHRI set for personality characterization (Celiktutan et al, 2017 ), and the EASE set for precise motion capturing (Meier et al, 2018 ). Secondly, data sets mimic the human perspective by holding objects in front of a perception device, such as a camera, to capture the diverse and complex but general characteristics of an environment setting, e.g., Core50 (Lomonaco and Maltoni, 2017 ), EMMI (Wang et al, 2017 ), and HOD-40 (Sun et al, 2018 ). And thirdly, humanoid robots are employed for establishing a data set, where multiple modalities are recorded in covering human-like action, i.e., including sensorimotor information, such as the MOD165 set (Nakamura and Nagai, 2017 ) and the Multimodal-HRI set (Azagra et al, 2017 ), or where multiple modalities are gathered from both robot and human in turn-table actions, like in the HARMONIC data set (Newman et al, 2018 ).…”
Section: Related Data Setsmentioning
confidence: 99%