2014 IEEE International Conference on Image Processing (ICIP) 2014
DOI: 10.1109/icip.2014.7026188
|View full text |Cite
|
Sign up to set email alerts
|

A dataset for Hand-Held Object Recognition

Abstract: Visual object recognition is just one of the many applications of camera-equipped smartphones. The ability to recognise objects through photos taken with wearable and handheld cameras is already possible through some of the larger internet search providers; yet, there is little rigorous analysis of the quality of search results, particularly where there is great disparity in image quality. This has motivated us to develop the Small Hand-held Object Recognition Test (SHORT). This includes a dataset that is suit… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 8 publications
(3 citation statements)
references
References 15 publications
(13 reference statements)
0
3
0
Order By: Relevance
“…There are few datasets available for this task. The majority of relevant datasets, such as HOD [28] and SHORT [29], are focused on single object detection, where each object is fully centered in the image as opposed to "in the wild" detection of handheld objects. In order to further research into handheld object detection, we created our own dataset with train, validation, and test splits based on data from the COCO-train 2017 dataset [30], an 80 class dataset of 118,287 images that is an industry-standard benchmark for object detection models.…”
Section: E Addendummentioning
confidence: 99%
“…There are few datasets available for this task. The majority of relevant datasets, such as HOD [28] and SHORT [29], are focused on single object detection, where each object is fully centered in the image as opposed to "in the wild" detection of handheld objects. In order to further research into handheld object detection, we created our own dataset with train, validation, and test splits based on data from the COCO-train 2017 dataset [30], an 80 class dataset of 118,287 images that is an industry-standard benchmark for object detection models.…”
Section: E Addendummentioning
confidence: 99%
“…Exemplary techniques are the well-known bar code or quick response (QR)-code, identifying a reference pattern on the floor, the image recognition of products by means of so-called planograms, cf. (Rivera-Rubio et al, 2014), or the transmission of coded light pulses above the critical flicker frequency of approximately 100 Hz. Using visual odometry, which correlates sequential images taken by the smart phone, a relative position estimate can be obtained without any knowledge of the environment, see (Shangguan et al, 2014).…”
Section: Sensors Actuators and Localization Techniquesmentioning
confidence: 99%
“…In total, more than 90,000 frames of video were labelled with positional ground-truth. The dataset is publicly available for download at http://rsm.bicv.org (Rivera-Rubio et al, 2014).…”
Section: The Datasetmentioning
confidence: 99%