Procedings of the British Machine Vision Conference 2017 2017
DOI: 10.5244/c.31.64
|View full text |Cite
|
Sign up to set email alerts
|

Dominant Set Clustering and Pooling for Multi-View 3D Object Recognition

Abstract: View based strategies for 3D object recognition have proven to be very successful. The state-of-the-art methods now achieve over 90% correct category level recognition performance on appearance images. We improve upon these methods by introducing a view clustering and pooling layer based on dominant sets. The key idea is to pool information from views which are similar and thus belong to the same cluster. The pooled feature vectors are then fed as inputs to the same layer, in a recurrent fashion. This recurren… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
124
0
1

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
4
1

Relationship

1
9

Authors

Journals

citations
Cited by 110 publications
(125 citation statements)
references
References 16 publications
(75 reference statements)
0
124
0
1
Order By: Relevance
“…Shapebased methods cover 3DShapeNets [26], VoxNet [27], Sub-VolSup [13], AniProbing [13], VRN & VRN-Ensemble [28], FPNN [29], PointNet [30], PointNet++ [31], O-CNN [32] and So-Net [33]. The view-based methods include MVCNN [12], DeepPano [34], PANORAMA-NN [35], PANORAMA-ENN [36], MVCNN-Alex [13], MVCNN-MultiRes [13], Pairwise [16], DomSetClust [14], RotationNet [17] Table 2 gives the detail comparison against state-of-theart discrete view-based methods with their different processing strategies. Apparently, RotationNet, DomSetClust and MVCNN-MultiRes get the better performance.…”
Section: Comparison With State-of-the-artsmentioning
confidence: 99%
“…Shapebased methods cover 3DShapeNets [26], VoxNet [27], Sub-VolSup [13], AniProbing [13], VRN & VRN-Ensemble [28], FPNN [29], PointNet [30], PointNet++ [31], O-CNN [32] and So-Net [33]. The view-based methods include MVCNN [12], DeepPano [34], PANORAMA-NN [35], PANORAMA-ENN [36], MVCNN-Alex [13], MVCNN-MultiRes [13], Pairwise [16], DomSetClust [14], RotationNet [17] Table 2 gives the detail comparison against state-of-theart discrete view-based methods with their different processing strategies. Apparently, RotationNet, DomSetClust and MVCNN-MultiRes get the better performance.…”
Section: Comparison With State-of-the-artsmentioning
confidence: 99%
“…High performance is achieved by taking advantage of the progress made in 2D classification by using a collection of 2D views of the object. They either pool over views [23], [24], apply more advanced schemes such as intelligently clustering before pooling [29] or jointly learning the corresponding object pose [13]. Beyond the power of such representations, they also take advantage of networks pretrained on large scale 2D datasets.…”
Section: A Geometric Deep Learningmentioning
confidence: 99%
“…The penalty λ of CorrReg is set as 5e −4 . We report in Table XII results of MvFusion-Net without or with CorrReg regularization, where we also compare with recent state-of-the-art results [73], [60], [61] on ModelNet40 whose multi-view images are prepared following the same style of 1 st camera set-up in [59] (i.e., 12 camera views pointing towards upright orientation of object models). Due to varying architectural designs, network optimizers, and feature aggregation schemes, results of different methods in Table XII may not be directly comparable; nevertheless, it confirms the efficacy of CorrReg for better feature learning and aggregation from multiple views of 3D object shapes.…”
Section: Multi-view Recognition Of 3d Object Shapesmentioning
confidence: 99%