2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00376
|View full text |Cite
|
Sign up to set email alerts
|

Category-Level Articulated Object Pose Estimation

Abstract: Human life is populated with articulated objects. Current Category-level Articulation Pose Estimation (CAPE) methods are studied under the singleinstance setting with a fixed kinematic structure for each category. Considering these limitations, we reform this problem setting for real-world environments and suggest a CAPE-Real (CAPER) task setting. This setting allows varied kinematic structures within a semantic category, and multiple instances to co-exist in an observation of real world. To support this task,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
154
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 159 publications
(159 citation statements)
references
References 22 publications
0
154
0
Order By: Relevance
“…Our proposed OMADNet can address the category-level articulated object pose estimation task given a single depth image as input. Overall, our work is closely related to [15] with significant different on methodology. Firstly, [15] takes a part-centered perspective, the joint states are estimated from the predicted part segmentations and poses.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Our proposed OMADNet can address the category-level articulated object pose estimation task given a single depth image as input. Overall, our work is closely related to [15] with significant different on methodology. Firstly, [15] takes a part-centered perspective, the joint states are estimated from the predicted part segmentations and poses.…”
Section: Related Workmentioning
confidence: 99%
“…Cateogry-level object pose estimation task aims at predicting poses of unseen objects in the same category, which is firstly introduced by NOCS [27]. A-NCSH [15] extends this task to articulated objects. Later, several works on this task are conducted on videos [11,16].…”
Section: Related Workmentioning
confidence: 99%
“…Such approaches first recover the position of semantic keypoints [56] in the images with neural networks, and then recover the 3D pose of the object by solving a geometric optimization problem [31,53,56,57,64]. In some works, a canonical coordinate space is predicted by a network instead of relying on geometric reasoning [14,22,41,78]. Lim et al [42] establish 2D-3D correspondences between images and textureless CAD models by using HOG descriptors, and render edgemaps of the CAD models.…”
Section: Related Workmentioning
confidence: 99%
“…Lu et al [47] proposed to train a probabilistic graphical model as a classifier to predict the appropriate grasp types (power grasp or precision grasp). Li et al [48] developed a deep network that uses a single depth point cloud to estimate the pose of an articulated object. Newbury et al [49] used two Convolutional Neural Networks (CNNs) to estimate both the placement rotations and stabilities and obtain the humanpreferred object placements and orientations.…”
Section: B Placement Estimationmentioning
confidence: 99%