2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2022
DOI: 10.1109/iros47612.2022.9981562
|View full text |Cite
|
Sign up to set email alerts
|

PoseIt: A Visual-Tactile Dataset of Holding Poses for Grasp Stability Analysis

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(4 citation statements)
references
References 30 publications
0
4
0
Order By: Relevance
“…The PoseIt Dataset [143] provides a comprehensive collection of multi-modal sensor data for predicting the stability of a grasp in a holding pose. This dataset includes tactile, RGB, and force/torque sensor readings collected during a sequence of timesteps, with a constant gripping force across all timesteps.…”
Section: Datasets Availablementioning
confidence: 99%
“…The PoseIt Dataset [143] provides a comprehensive collection of multi-modal sensor data for predicting the stability of a grasp in a holding pose. This dataset includes tactile, RGB, and force/torque sensor readings collected during a sequence of timesteps, with a constant gripping force across all timesteps.…”
Section: Datasets Availablementioning
confidence: 99%
“…Cui et al [5] utilized a 3D CNN-based visual-tactile fusion network to evaluate the grasp state of deformable objects. Kanitkar et al [7] presented a multimodal dataset consisting of visual-tactile information to investigate the impact of varied holding poses on grasp stability. Nevertheless, the adoption of simplistic feature-level fusion approaches in these studies resulted in a restricted exploitation of complementary information and an inability to effectively capture the interplay among unimodal features.…”
Section: A Grasp Stability Evaluationmentioning
confidence: 99%
“…𝐗 β„Ž ← 𝐗 β„Ž + MHSA(𝐗 β„Ž + 𝐏 β„Ž , 𝐗 𝑣 + 𝐏 𝑣 , 𝐗 𝑣 ) (7) where 𝐗 𝑣 and 𝐗 β„Ž denote the feature sequences of the visual and tactile channels, respectively.…”
Section: Feature Extractionmentioning
confidence: 99%
See 1 more Smart Citation