2014 12th International Workshop on Content-Based Multimedia Indexing (CBMI) 2014
DOI: 10.1109/cbmi.2014.6849835
|View full text |Cite
|
Sign up to set email alerts
|

Evaluation of second-order visual features for land-use classification

Abstract: Abstract-This paper investigates the use of recent visual features based on second-order statistics, as well as new processing techniques to improve the quality of features. More specifically, we present and evaluate Fisher Vectors (FV), Vectors of Locally Aggregated Descriptors (VLAD), and Vectors of Locally Aggregated Tensors (VLAT). These techniques are combined with several normalization techniques, such as power law normalization and orthogonalisation/whitening of descriptor spaces. Results on the UC Merc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
34
0

Year Published

2015
2015
2023
2023

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 70 publications
(34 citation statements)
references
References 19 publications
0
34
0
Order By: Relevance
“…2010 SPM [1] 74.00 2010 SPCK++ [1] 76.05 2015 Saliency-UFL [2] 82.72±1.18 2014 Bag-of-SIFT [3] 85.37±1.56 Single-view deep learning 88.00±2.88 2014 SAL-LDA [5] 88.33 2015 Pyramid of spatial relatons [6] 89.1 2014 UFL [3] 90.26±1.51 Multiview deep learning 93.48±0.82 2014 VLAT [7] 94.3…”
Section: Date Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…2010 SPM [1] 74.00 2010 SPCK++ [1] 76.05 2015 Saliency-UFL [2] 82.72±1.18 2014 Bag-of-SIFT [3] 85.37±1.56 Single-view deep learning 88.00±2.88 2014 SAL-LDA [5] 88.33 2015 Pyramid of spatial relatons [6] 89.1 2014 UFL [3] 90.26±1.51 Multiview deep learning 93.48±0.82 2014 VLAT [7] 94.3…”
Section: Date Methodsmentioning
confidence: 99%
“…The highest accuracies for the UC Merced dataset have been achieved with unsupervised feature learning (UFL) [3] and the vector of locally aggregated tensors (VLAT) method [7], which is an extension of visual dictionary approaches like bag-of-words. Single-view DCNN is outperformed by these methods, but the 90.26% accuracy of UFL can be improved upon with a multiview DCNN which achieves 93.48%.…”
Section: Accuracy Comparisonmentioning
confidence: 99%
See 2 more Smart Citations
“…The current state-of-the-art for this dataset is 94.3% (Negrel et al, 2014) acquired by aggregating tensor products of local descriptors.…”
Section: Datasetsmentioning
confidence: 99%