2020
DOI: 10.3390/s21010113
|View full text |Cite
|
Sign up to set email alerts
|

Transfer of Learning from Vision to Touch: A Hybrid Deep Convolutional Neural Network for Visuo-Tactile 3D Object Recognition

Abstract: Transfer of learning or leveraging a pre-trained network and fine-tuning it to perform new tasks has been successfully applied in a variety of machine intelligence fields, including computer vision, natural language processing and audio/speech recognition. Drawing inspiration from neuroscience research that suggests that both visual and tactile stimuli rouse similar neural networks in the human brain, in this work, we explore the idea of transferring learning from vision to touch in the context of 3D object re… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(8 citation statements)
references
References 28 publications
0
8
0
Order By: Relevance
“…The related work in object recognition by handling is quite diverse due to developments in several aspects of this subject, beginning with studies of human haptic behavior [7] from physiological and psychological studies to understand human haptic perception and unfamiliar objects surveys for recognition [8]. Other studies include humanoid robot hand system design to enable object handling [3], [4], [9], [10], control of the robot hand system of a humanoid robot [11], [12], tactile sensor array design for humanoids robot hand [5], [6], [13], [14], [15], [16], [17], [18], [19], [20], [21], [22], [23], [24], [25], [26], [27], [28], [29], [30], [31], [32]. Methods for recognizing objects based on tactile image recognition [6], [15], [18], [20], [21], [22], [23], [24], [28], [29], [31], [32], [33], [34], [35],…”
Section: Theory and Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The related work in object recognition by handling is quite diverse due to developments in several aspects of this subject, beginning with studies of human haptic behavior [7] from physiological and psychological studies to understand human haptic perception and unfamiliar objects surveys for recognition [8]. Other studies include humanoid robot hand system design to enable object handling [3], [4], [9], [10], control of the robot hand system of a humanoid robot [11], [12], tactile sensor array design for humanoids robot hand [5], [6], [13], [14], [15], [16], [17], [18], [19], [20], [21], [22], [23], [24], [25], [26], [27], [28], [29], [30], [31], [32]. Methods for recognizing objects based on tactile image recognition [6], [15], [18], [20], [21], [22], [23], [24], [28], [29], [31], [32], [33], [34], [35],…”
Section: Theory and Related Workmentioning
confidence: 99%
“…Other studies include humanoid robot hand system design to enable object handling [3], [4], [9], [10], control of the robot hand system of a humanoid robot [11], [12], tactile sensor array design for humanoids robot hand [5], [6], [13], [14], [15], [16], [17], [18], [19], [20], [21], [22], [23], [24], [25], [26], [27], [28], [29], [30], [31], [32]. Methods for recognizing objects based on tactile image recognition [6], [15], [18], [20], [21], [22], [23], [24], [28], [29], [31], [32], [33], [34], [35], [36], [37], [38], [39], [40], [4...…”
Section: Theory and Related Workmentioning
confidence: 99%
“…Bednarek et al [145] conducted grasp classification experiments on the BiGS dataset to compare the performance of four multimodal fusion algorithms of late fusion, MoE, intermediate fusion and LMF. Rouhafzay et al [146] retrained the convolutional neural network on the successful cases of the BiGS dataset and proposed a hybrid framework MobileNetV2. Results proved that their pretrained deep convolutional neural network on visual images could be effectively transferred to the tactile dataset for classification tasks.…”
Section: B Datasetsmentioning
confidence: 99%
“…The dataset was established to first fuse and share the visual and tactile characteristics of different fabrics and then to improve the accuracy of fabric texture recognition tasks. Rouhafzay et al [146] selected 12 kinds of tactile data from the ViTac dataset to retrain and fine-tune their pretrained deep convolutional neural network to ensure the quality of transfer learning. Lee et al [147] proposed a cross-modal sensory data generating framework using a conditional generative adversarial network to generate pseudovisual data from tactile data or to generate pseudotactile data from visual data.…”
Section: B Datasetsmentioning
confidence: 99%
“…By establishing cross‐mode association between vision and tactile measurement, the knowledge obtained from one modality can be transferred to the other, consequently improving or extending the sensory capabilities. [ 154,155 ]…”
Section: Vision‐ and Touch‐enabled Humanoidsmentioning
confidence: 99%